You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
"""Config for the tokenizer pool. If None, will use synchronous
1542
+
tokenization."""
1460
1543
1461
-
# Whether to profile Ray workers with nsight, see https://docs.ray.io/en/latest/ray-observability/user-guides/profiling.html#profiling-nsight-profiler.
1462
1544
ray_workers_use_nsight: bool=False
1545
+
"""Whether to profile Ray workers with nsight, see https://docs.ray.io/en/latest/ray-observability/user-guides/profiling.html#profiling-nsight-profiler."""
1463
1546
1464
-
# ray distributed model workers placement group.
1465
1547
placement_group: Optional["PlacementGroup"] =None
1548
+
"""ray distributed model workers placement group."""
1466
1549
1467
-
# Backend to use for distributed model
1468
-
# workers, either "ray" or "mp" (multiprocessing). If the product
1469
-
# of pipeline_parallel_size and tensor_parallel_size is less than
1470
-
# or equal to the number of GPUs available, "mp" will be used to
1471
-
# keep processing on a single host. Otherwise, this will default
1472
-
# to "ray" if Ray is installed and fail otherwise. Note that tpu
1473
-
# and hpu only support Ray for distributed inference.
1474
1550
distributed_executor_backend: Optional[Union[str,
1475
1551
type["ExecutorBase"]]] =None
1552
+
"""Backend to use for distributed model
1553
+
workers, either "ray" or "mp" (multiprocessing). If the product
1554
+
of pipeline_parallel_size and tensor_parallel_size is less than
1555
+
or equal to the number of GPUs available, "mp" will be used to
1556
+
keep processing on a single host. Otherwise, this will default
1557
+
to "ray" if Ray is installed and fail otherwise. Note that tpu
1558
+
and hpu only support Ray for distributed inference."""
1476
1559
1477
-
# the full name of the worker class to use. If "auto", the worker class
1478
-
# will be determined based on the platform.
1479
1560
worker_cls: str="auto"
1561
+
"""The full name of the worker class to use. If "auto", the worker class
1562
+
will be determined based on the platform."""
1480
1563
sd_worker_cls: str="auto"
1564
+
"""The full name of the worker class to use for speculative decofing.
1565
+
If "auto", the worker class will be determined based on the platform."""
1481
1566
worker_extension_cls: str=""
1567
+
"""The full name of the worker extension class to use. The worker extension
1568
+
class is dynamically inherited by the worker class. This is used to inject
1569
+
new attributes and methods to the worker class for use in collective_rpc
1570
+
calls."""
1482
1571
1483
-
# world_size is TPxPP, it affects the number of workers we create.
1484
1572
world_size: int=field(init=False)
1485
-
# world_size_across_dp is TPxPPxDP, it is the size of the world
1486
-
# including data parallelism.
1573
+
"""world_size is TPxPP, it affects the number of workers we create."""
1487
1574
world_size_across_dp: int=field(init=False)
1575
+
"""world_size_across_dp is TPxPPxDP, it is the size of the world
0 commit comments