File tree Expand file tree Collapse file tree 1 file changed +4
-4
lines changed
Expand file tree Collapse file tree 1 file changed +4
-4
lines changed Original file line number Diff line number Diff line change @@ -47,10 +47,10 @@ def test_llama_7b_lora_tp2():
4747 kv_cache_config = global_kv_cache_config )
4848
4949
50- @pytest .mark .gpu2
51- @pytest . mark . skip ( reason = " https://nvbugs/5682551" )
50+ @pytest .mark .gpu4
51+ @skip_ray # https://nvbugs/5682551
5252@test_lora_with_and_without_cuda_graph
53- def test_llama_7b_multi_lora_tp2 (cuda_graph_config ):
53+ def test_llama_7b_multi_lora_tp4 (cuda_graph_config ):
5454 # For LoRA checkpoints without finetuned embedding and lm_head, we can either:
5555 # (1) specify lora_target_modules, or
5656 # (2) provide a lora_dir to infer the lora_target_modules.
@@ -61,7 +61,7 @@ def test_llama_7b_multi_lora_tp2(cuda_graph_config):
6161 check_llama_7b_multi_lora_from_request_test_harness (
6262 LLM ,
6363 lora_config = lora_config ,
64- tensor_parallel_size = 2 ,
64+ tensor_parallel_size = 4 ,
6565 kv_cache_config = global_kv_cache_config ,
6666 cuda_graph_config = cuda_graph_config )
6767
You can’t perform that action at this time.
0 commit comments