Skip to content

Commit 5ddaccf

Browse files
committed
Use tp4 in test; use skip ray
Signed-off-by: Jiayu Chang <jiayuc@nvidia.com>
1 parent 5b02b7a commit 5ddaccf

File tree

1 file changed

+4
-4
lines changed

1 file changed

+4
-4
lines changed

tests/unittest/llmapi/test_llm_multi_gpu_pytorch.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -47,10 +47,10 @@ def test_llama_7b_lora_tp2():
4747
kv_cache_config=global_kv_cache_config)
4848

4949

50-
@pytest.mark.gpu2
51-
@pytest.mark.skip(reason="https://nvbugs/5682551")
50+
@pytest.mark.gpu4
51+
@skip_ray # https://nvbugs/5682551
5252
@test_lora_with_and_without_cuda_graph
53-
def test_llama_7b_multi_lora_tp2(cuda_graph_config):
53+
def test_llama_7b_multi_lora_tp4(cuda_graph_config):
5454
# For LoRA checkpoints without finetuned embedding and lm_head, we can either:
5555
# (1) specify lora_target_modules, or
5656
# (2) provide a lora_dir to infer the lora_target_modules.
@@ -61,7 +61,7 @@ def test_llama_7b_multi_lora_tp2(cuda_graph_config):
6161
check_llama_7b_multi_lora_from_request_test_harness(
6262
LLM,
6363
lora_config=lora_config,
64-
tensor_parallel_size=2,
64+
tensor_parallel_size=4,
6565
kv_cache_config=global_kv_cache_config,
6666
cuda_graph_config=cuda_graph_config)
6767

0 commit comments

Comments
 (0)