We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 2e4c153 commit 2d27c1aCopy full SHA for 2d27c1a
_unittests/ut_torch_export_patches/test_patch_transformers.py
@@ -461,14 +461,16 @@ def forward(
461
target_opset=22,
462
)
463
# exporter_kwargs={"report":True} if exporter != "custom" else {}
464
- self.assert_onnx_disc(
465
- f"test_patched_qwen2_5_vl_vision_attention_forward-{exporter}",
466
- onnx.load(filename),
467
- instance,
468
- inputs,
469
- atol=1e-3,
470
- rtol=1,
471
- )
+ if torch.cuda.is_available():
+ self.assert_onnx_disc(
+ f"test_patched_qwen2_5_vl_vision_attention_forward-{exporter}",
+ onnx.load(filename),
+ instance,
+ inputs,
+ atol=1e-3,
+ rtol=1,
472
+ providers=["CUDAExecutionProvider"],
473
+ )
474
self.clean_dump()
475
476
@requires_transformers("4.99")
0 commit comments