Skip to content

Commit 57b9f02

Browse files
authored
[Bugfix] Fix disaggregated pd error (#2242)
### What this PR does / why we need it? Fix `ascend_env has no attr VLLM_ASCEND_ENABLE_CHUNK_MC2`, remove useless lines - vLLM version: v0.10.0 - vLLM main: vllm-project/vllm@9edd1db --------- Signed-off-by: wangli <[email protected]>
1 parent 26fc36b commit 57b9f02

File tree

1 file changed

+0
-16
lines changed

1 file changed

+0
-16
lines changed

vllm_ascend/worker/model_runner_v1.py

Lines changed: 0 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -646,22 +646,6 @@ def _get_forward_metadata_across_dp_and_pad(
646646
return maybe_padded_num_tokens, None, with_prefill, enable_dbo
647647
return num_tokens, None, with_prefill, enable_dbo
648648

649-
if self.is_kv_producer and not envs_ascend.VLLM_ASCEND_ENABLE_CHUNK_MC2:
650-
num_tokens_across_dp = torch.tensor([num_tokens] * self.dp_size,
651-
device="cpu",
652-
dtype=torch.int32)
653-
return num_tokens, num_tokens_across_dp, True, enable_dbo
654-
655-
if self.is_kv_consumer and self.torchair_graph_enabled and len(
656-
self.torchair_graph_batch_sizes
657-
) == 1 and not self.in_profile_run:
658-
max_num_decode_tokens = self.torchair_graph_batch_sizes[0]
659-
num_tokens_across_dp = torch.tensor([max_num_decode_tokens] *
660-
self.dp_size,
661-
device="cpu",
662-
dtype=torch.int32)
663-
return max_num_decode_tokens, num_tokens_across_dp, False, enable_dbo
664-
665649
maybe_padded_num_tokens = num_tokens
666650
num_tokens_across_dp, with_prefill, enable_dbo = self._get_forward_metadata_across_dp(
667651
num_tokens, with_prefill, enable_dbo)

0 commit comments

Comments
 (0)