Skip to content

Commit a355755

Browse files
committed
fix
Signed-off-by: Wuxun Zhang <[email protected]>
1 parent 5ad7ff8 commit a355755

File tree

2 files changed

+0
-7
lines changed

2 files changed

+0
-7
lines changed

vllm_gaudi/distributed/device_communicators/hpu_communicator.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -65,9 +65,6 @@ def all_gather(self, input_: torch.Tensor, dim: int = -1) -> torch.Tensor:
6565
def dispatch(
6666
self, hidden_states: torch.Tensor,
6767
router_logits: torch.Tensor) -> tuple[torch.Tensor, torch.Tensor]:
68-
"""
69-
all-gather based dispatch for HPUCommunicator.
70-
"""
7168
cu_tokens_across_dp_cpu = get_forward_context(
7269
).dp_metadata.cu_tokens_across_dp_cpu
7370
hidden_states_across_dp = naive_multicast(hidden_states,

vllm_gaudi/v1/worker/hpu_model_runner.py

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1229,7 +1229,6 @@ def _form_prefill_batch(self, contents):
12291229

12301230
query_lens = _async_h2d_tensor(query_lens, torch.int32)
12311231
token_ids = _async_h2d_tensor(token_ids, torch.int32)
1232-
12331232
token_positions = _async_h2d_tensor(token_positions, torch.int32)
12341233
token_slots = _async_h2d_tensor(token_slots, torch.int64)
12351234
logits_indices = _async_h2d_tensor(logits_indices, torch.int32)
@@ -1296,7 +1295,6 @@ def _prepare_decode_inputs(self, num_decodes,
12961295
num_decodes, sum(num_blocks))[0]
12971296

12981297
# # dp aware padding
1299-
assert padded_batch_size is not None
13001298
padded_batch_size += self.get_dp_padding(padded_batch_size)
13011299

13021300
block_tables_list = []
@@ -1754,8 +1752,6 @@ def execute_model(self,
17541752

17551753
######################### PREFILLS #########################
17561754
if num_prefills > 0:
1757-
# Wuxun: merged prefill forward if enabled
1758-
# 2D bucketing or merged prefill bucketing
17591755
htorch.core.mark_step()
17601756
for idx, (req_id, prompt_len, token_ids, position_ids,
17611757
attn_metadata, logits_indices,

0 commit comments

Comments
 (0)