Skip to content

Commit 0c54343

Browse files
committed
remove enable_async_comms
Signed-off-by: Sage Moore <[email protected]>
1 parent 4718a2d commit 0c54343

File tree

4 files changed

+1
-14
lines changed

4 files changed

+1
-14
lines changed

vllm/compilation/ubatch_wrapper.py

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -209,9 +209,7 @@ def _make_ubatch_metadata(self, ubatch_slices, attn_metadata, input_ids,
209209
compute_stream=compute_stream,
210210
forward_contexts=forward_contexts,
211211
ready_barrier=self.ready_barrier,
212-
device=self.device,
213-
enable_async_comms=self.vllm_config.parallel_config.
214-
enable_async_comms)
212+
device=self.device)
215213

216214
ubatch_metadata: list[UbatchMetadata] = []
217215
for i, ubatch_slice in enumerate(ubatch_slices):

vllm/config/parallel.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -135,9 +135,6 @@ class ParallelConfig:
135135
request is greater than this threshold, microbatching will be used.
136136
Otherwise, the request will be processed in a single batch."""
137137

138-
enable_async_comms: bool = False
139-
"""enable async comms"""
140-
141138
ray_workers_use_nsight: bool = False
142139
"""Whether to profile Ray workers with nsight, see https://docs.ray.io/en/latest/ray-observability/user-guides/profiling.html#profiling-nsight-profiler."""
143140

vllm/engine/arg_utils.py

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -317,7 +317,6 @@ class EngineArgs:
317317
enable_expert_parallel: bool = ParallelConfig.enable_expert_parallel
318318
enable_microbatching: bool = ParallelConfig.enable_microbatching
319319
microbatching_token_threshold: int = ParallelConfig.microbatching_token_threshold
320-
enable_async_comms: bool = ParallelConfig.enable_async_comms
321320
eplb_config: EPLBConfig = get_field(ParallelConfig, "eplb_config")
322321
enable_eplb: bool = ParallelConfig.enable_eplb
323322
num_redundant_experts: int = EPLBConfig.num_redundant_experts
@@ -683,8 +682,6 @@ def add_cli_args(parser: FlexibleArgumentParser) -> FlexibleArgumentParser:
683682
**parallel_kwargs["enable_microbatching"])
684683
parallel_group.add_argument("--microbatching-token-threshold",
685684
**parallel_kwargs["microbatching_token_threshold"])
686-
parallel_group.add_argument("--enable-async-comms",
687-
**parallel_kwargs["enable_async_comms"])
688685
parallel_group.add_argument("--enable-eplb",
689686
**parallel_kwargs["enable_eplb"])
690687
parallel_group.add_argument("--eplb-config",
@@ -1307,7 +1304,6 @@ def create_engine_config(
13071304
enable_expert_parallel=self.enable_expert_parallel,
13081305
enable_microbatching=self.enable_microbatching,
13091306
microbatching_token_threshold=self.microbatching_token_threshold,
1310-
enable_async_comms=self.enable_async_comms,
13111307
enable_eplb=self.enable_eplb,
13121308
eplb_config=self.eplb_config,
13131309
max_parallel_loading_workers=self.max_parallel_loading_workers,

vllm/v1/worker/ubatching.py

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,6 @@ def __init__(self,
2727
cpu_signal_event: threading.Event,
2828
gpu_comm_done_event: torch.cuda.Event,
2929
gpu_compute_done_event: torch.cuda.Event,
30-
enable_async_comms: bool,
3130
schedule: str = "default"):
3231
self.id = id
3332
self.comm_stream = comm_stream
@@ -39,7 +38,6 @@ def __init__(self,
3938
self.current_stream = compute_stream
4039
self.gpu_comm_done_event = gpu_comm_done_event
4140
self.gpu_compute_done_event = gpu_compute_done_event
42-
self.enable_async_comms = enable_async_comms
4341
self.schedule = schedule
4442
self.recv_hook = None
4543

@@ -175,7 +173,6 @@ def make_ubatch_contexts(
175173
forward_contexts: list[ForwardContext],
176174
ready_barrier: threading.Barrier,
177175
device: Optional[torch.device] = None,
178-
enable_async_comms: bool = False,
179176
schedule: str = "default",
180177
) -> list[UBatchContext]:
181178
assert num_micro_batches == 2, "only been tested with 2 micro-batches"
@@ -206,7 +203,6 @@ def make_ubatch_contexts(
206203
num_micro_batches],
207204
gpu_comm_done_event=gpu_comm_done_events[i],
208205
gpu_compute_done_event=gpu_compute_done_events[i],
209-
enable_async_comms=enable_async_comms,
210206
schedule=schedule)
211207
ctxs.append(ctx)
212208

0 commit comments

Comments
 (0)