Skip to content

Commit 46d5570

Browse files
committed
Add ray stage with 4 h100 gpus to CI and fix sampled logprobs in TRTLLM sampler
Signed-off-by: Shuyi Xiong <219646547+shuyixiong@users.noreply.github.com>
1 parent fcda1a1 commit 46d5570

File tree

3 files changed

+3
-1
lines changed

3 files changed

+3
-1
lines changed

jenkins/L0_MergeRequest.groovy

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -734,6 +734,7 @@ def getMultiGpuFileChanged(pipeline, testFilter, globalVars)
734734
"tests/unittest/_torch/auto_deploy/unit/multigpu",
735735
"tests/unittest/_torch/multi_gpu/",
736736
"tests/unittest/_torch/multi_gpu_modeling/",
737+
"tests/unittest/_torch/ray_orchestrator/multi_gpu/",
737738
"tests/unittest/disaggregated/",
738739
"tests/unittest/llmapi/test_llm_multi_gpu.py",
739740
"tests/unittest/llmapi/test_llm_multi_gpu_pytorch.py",

jenkins/L0_Test.groovy

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -2938,6 +2938,7 @@ def launchTestJobs(pipeline, testFilter)
29382938
"DGX_H100-4_GPUs-PyTorch-DeepSeek-1": ["dgx-h100-x4-oci", "l0_dgx_h100", 1, 1, 4],
29392939
"DGX_H100-4_GPUs-PyTorch-GptOss-1": ["dgx-h100-x4-oci", "l0_dgx_h100", 1, 1, 4],
29402940
"DGX_H100-4_GPUs-PyTorch-Others-1": ["dgx-h100-x4-oci", "l0_dgx_h100", 1, 1, 4],
2941+
"DGX_H100-4_GPUs-PyTorch-Ray-1": ["dgx-h100-x4-oci", "l0_dgx_h100", 1, 1, 4],
29412942
"B300-PyTorch-1": ["b300-single", "l0_b300", 1, 1],
29422943
"DGX_B200-4_GPUs-PyTorch-1": ["b200-x4", "l0_dgx_b200", 1, 2, 4],
29432944
"DGX_B200-4_GPUs-PyTorch-2": ["b200-x4", "l0_dgx_b200", 2, 2, 4],

tensorrt_llm/_torch/pyexecutor/sampler.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3049,7 +3049,7 @@ def update_requests_single_beam_single_step(self, state: SampleStateTRTLLM):
30493049

30503050
log_probs_host = log_probs_host_tensor[seq_slots, 0, seq_lens].tolist()
30513051
idx = 0
3052-
for request in reqs_with_new_tokens:
3052+
for seq_slot, request in zip(seq_slots, reqs_with_new_tokens):
30533053
if request.py_return_log_probs:
30543054
log_probs = [
30553055
{

0 commit comments

Comments
 (0)