File tree Expand file tree Collapse file tree 1 file changed +2
-2
lines changed
tests/e2e/patched/lora_kernels Expand file tree Collapse file tree 1 file changed +2
-2
lines changed Original file line number Diff line number Diff line change @@ -144,7 +144,7 @@ def test_swiglu_mlp_integration(small_llama_model):
144
144
def test_geglu_model_integration ():
145
145
"""Test GeGLU activation with Gemma model."""
146
146
model = AutoModelForCausalLM .from_pretrained (
147
- "mhenrichsen/gemma-2b" , torch_dtype = torch .float16 , device_map = "auto "
147
+ "mhenrichsen/gemma-2b" , torch_dtype = torch .float16 , device_map = "cuda:0 "
148
148
)
149
149
peft_config = get_peft_config (
150
150
{
@@ -347,7 +347,7 @@ def test_model_architecture(model_config):
347
347
"""Test LoRA kernel patches across different model architectures."""
348
348
# Load model with appropriate dtype
349
349
model = AutoModelForCausalLM .from_pretrained (
350
- model_config ["name" ], torch_dtype = model_config ["dtype" ], device_map = "auto "
350
+ model_config ["name" ], torch_dtype = model_config ["dtype" ], device_map = "cuda:0 "
351
351
)
352
352
353
353
# Apply LoRA configuration
You can’t perform that action at this time.
0 commit comments