@@ -309,6 +309,7 @@ def is_inplace(op, variant):
309
309
skip ('pca_lowrank' , '' ), # fails on cuda, runs okay on cpu
310
310
skip ('svd_lowrank' , '' ), # fails on cuda, runs okay on cpu
311
311
skip ('nn.functional.dropout2d' , '' ), # fails on cuda, runs okay on cpu
312
+ xfail ('__getitem__' , device_type = 'cuda' ),
312
313
}
313
314
314
315
@@ -317,6 +318,18 @@ class TestOperators(TestCase):
317
318
@skipOps ('TestOperators' , 'test_grad' , vjp_fail .union ({
318
319
skip ('nn.functional.fractional_max_pool2d' ), # fails on cuda, runs okay on cpu
319
320
skip ('nn.functional.fractional_max_pool3d' ), # fails on cuda, runs okay on cpu
321
+ xfail ('__getitem__' , 'functorch' , device_type = 'cuda' ),
322
+ xfail ('_masked.amax' , device_type = 'cuda' ),
323
+ xfail ('_masked.amin' , device_type = 'cuda' ),
324
+ xfail ('_masked.log_softmax' , device_type = 'cuda' ),
325
+ xfail ('_masked.mean' , device_type = 'cuda' ),
326
+ xfail ('_masked.norm' , device_type = 'cuda' ),
327
+ xfail ('_masked.prod' , device_type = 'cuda' ),
328
+ xfail ('_masked.softmax' , device_type = 'cuda' ),
329
+ xfail ('_masked.softmin' , device_type = 'cuda' ),
330
+ xfail ('_masked.std' , device_type = 'cuda' ),
331
+ xfail ('_masked.sum' , device_type = 'cuda' ),
332
+ xfail ('_masked.var' , device_type = 'cuda' ),
320
333
}))
321
334
@opsToleranceOverride ('TestOperators' , 'test_grad' , (
322
335
tol1 ('nn.functional.binary_cross_entropy_with_logits' ,
@@ -396,6 +409,16 @@ def wrapped_fn(*args, **kwargs):
396
409
skip ('nn.functional.max_unpool1d' ), # fails everywhere except on mac
397
410
skip ('nn.functional.max_unpool2d' ), # fails everywhere except on windows
398
411
xfail ('nn.functional.max_unpool3d' ),
412
+ xfail ('__getitem__' , device_type = 'cuda' ),
413
+ xfail ('_masked.log_softmax' , device_type = 'cuda' ),
414
+ xfail ('_masked.mean' , device_type = 'cuda' ),
415
+ xfail ('_masked.norm' , device_type = 'cuda' ),
416
+ xfail ('_masked.prod' , device_type = 'cuda' ),
417
+ xfail ('_masked.softmax' , device_type = 'cuda' ),
418
+ xfail ('_masked.softmin' , device_type = 'cuda' ),
419
+ xfail ('_masked.std' , device_type = 'cuda' ),
420
+ xfail ('_masked.sum' , device_type = 'cuda' ),
421
+ xfail ('_masked.var' , device_type = 'cuda' ),
399
422
}))
400
423
@opsToleranceOverride ('TestOperators' , 'test_jvp' , (
401
424
tol1 ('nn.functional.conv_transpose3d' ,
@@ -443,6 +466,19 @@ def test_jvp(self, device, dtype, op):
443
466
xfail ('nn.functional.dropout2d' , '' ),
444
467
xfail ('nn.functional.feature_alpha_dropout' , 'without_train' ),
445
468
xfail ('svd_lowrank' , '' ),
469
+
470
+ xfail ('__getitem__' , 'functorch' , device_type = 'cuda' ),
471
+ xfail ('_masked.amax' , device_type = 'cuda' ),
472
+ xfail ('_masked.amin' , device_type = 'cuda' ),
473
+ xfail ('_masked.log_softmax' , device_type = 'cuda' ),
474
+ xfail ('_masked.mean' , device_type = 'cuda' ),
475
+ xfail ('_masked.norm' , device_type = 'cuda' ),
476
+ xfail ('_masked.prod' , device_type = 'cuda' ),
477
+ xfail ('_masked.softmax' , device_type = 'cuda' ),
478
+ xfail ('_masked.softmin' , device_type = 'cuda' ),
479
+ xfail ('_masked.std' , device_type = 'cuda' ),
480
+ xfail ('_masked.sum' , device_type = 'cuda' ),
481
+ xfail ('_masked.var' , device_type = 'cuda' ),
446
482
}))
447
483
@opsToleranceOverride ('TestOperators' , 'test_vjp' , (
448
484
tol1 ('nn.functional.conv_transpose3d' ,
@@ -488,6 +524,19 @@ def _test(_op):
488
524
skip ('nn.functional.fractional_max_pool2d' ), # randomness
489
525
skip ('nn.functional.fractional_max_pool3d' ), # randomness
490
526
xfail ('nn.functional.binary_cross_entropy' ), # testing problem
527
+
528
+ xfail ('__getitem__' , 'functorch' , device_type = 'cuda' ),
529
+ xfail ('_masked.amax' , device_type = 'cuda' ),
530
+ xfail ('_masked.amin' , device_type = 'cuda' ),
531
+ xfail ('_masked.log_softmax' , device_type = 'cuda' ),
532
+ xfail ('_masked.mean' , device_type = 'cuda' ),
533
+ xfail ('_masked.norm' , device_type = 'cuda' ),
534
+ xfail ('_masked.prod' , device_type = 'cuda' ),
535
+ xfail ('_masked.softmax' , device_type = 'cuda' ),
536
+ xfail ('_masked.softmin' , device_type = 'cuda' ),
537
+ xfail ('_masked.std' , device_type = 'cuda' ),
538
+ xfail ('_masked.sum' , device_type = 'cuda' ),
539
+ xfail ('_masked.var' , device_type = 'cuda' ),
491
540
}))
492
541
@opsToleranceOverride ('TestOperators' , 'test_vjpvjp' , (
493
542
tol1 ('nn.functional.conv_transpose3d' ,
@@ -623,6 +672,19 @@ def vjp_of_vjp(*args_and_cotangents):
623
672
# NYI: querying is_contiguous inside of vmap for memory_format other than torch.contiguous_format
624
673
xfail ('nn.functional.max_unpool2d' ),
625
674
xfail ('nn.functional.max_unpool2d' , 'grad' ),
675
+
676
+ xfail ('__getitem__' , 'functorch' , device_type = 'cuda' ),
677
+ xfail ('_masked.amax' , device_type = 'cuda' ),
678
+ xfail ('_masked.amin' , device_type = 'cuda' ),
679
+ xfail ('_masked.log_softmax' , device_type = 'cuda' ),
680
+ xfail ('_masked.mean' , device_type = 'cuda' ),
681
+ xfail ('_masked.norm' , device_type = 'cuda' ),
682
+ xfail ('_masked.prod' , device_type = 'cuda' ),
683
+ xfail ('_masked.softmax' , device_type = 'cuda' ),
684
+ xfail ('_masked.softmin' , device_type = 'cuda' ),
685
+ xfail ('_masked.std' , device_type = 'cuda' ),
686
+ xfail ('_masked.sum' , device_type = 'cuda' ),
687
+ xfail ('_masked.var' , device_type = 'cuda' ),
626
688
})
627
689
628
690
@ops (functorch_lagging_op_db + additional_op_db , allowed_dtypes = (torch .float ,))
@@ -711,6 +773,19 @@ def test_vmapvjp(self, device, dtype, op):
711
773
xfail ('nn.functional.max_unpool1d' , device_type = 'cpu' ),
712
774
xfail ('nn.functional.max_unpool2d' ),
713
775
xfail ('nn.functional.max_unpool3d' ),
776
+
777
+ xfail ('__getitem__' , device_type = 'cuda' ),
778
+ xfail ('_masked.amax' , device_type = 'cuda' ),
779
+ xfail ('_masked.amin' , device_type = 'cuda' ),
780
+ xfail ('_masked.log_softmax' , device_type = 'cuda' ),
781
+ xfail ('_masked.mean' , device_type = 'cuda' ),
782
+ xfail ('_masked.norm' , device_type = 'cuda' ),
783
+ xfail ('_masked.prod' , device_type = 'cuda' ),
784
+ xfail ('_masked.softmax' , device_type = 'cuda' ),
785
+ xfail ('_masked.softmin' , device_type = 'cuda' ),
786
+ xfail ('_masked.std' , device_type = 'cuda' ),
787
+ xfail ('_masked.sum' , device_type = 'cuda' ),
788
+ xfail ('_masked.var' , device_type = 'cuda' ),
714
789
})
715
790
def test_vmapjvp (self , device , dtype , op ):
716
791
if is_inplace (op , op .get_op ()):
@@ -787,6 +862,19 @@ def test_vmapjvp(self, device, dtype, op):
787
862
# BUG: runs and produces numerical differences
788
863
xfail ('nn.functional.max_unpool2d' ),
789
864
xfail ('nn.functional.max_unpool3d' ),
865
+
866
+ xfail ('__getitem__' , device_type = 'cuda' ),
867
+ xfail ('_masked.amax' , device_type = 'cuda' ),
868
+ xfail ('_masked.amin' , device_type = 'cuda' ),
869
+ xfail ('_masked.log_softmax' , device_type = 'cuda' ),
870
+ xfail ('_masked.mean' , device_type = 'cuda' ),
871
+ xfail ('_masked.norm' , device_type = 'cuda' ),
872
+ xfail ('_masked.prod' , device_type = 'cuda' ),
873
+ xfail ('_masked.softmax' , device_type = 'cuda' ),
874
+ xfail ('_masked.softmin' , device_type = 'cuda' ),
875
+ xfail ('_masked.std' , device_type = 'cuda' ),
876
+ xfail ('_masked.sum' , device_type = 'cuda' ),
877
+ xfail ('_masked.var' , device_type = 'cuda' ),
790
878
}
791
879
792
880
@ops (functorch_lagging_op_db , allowed_dtypes = (torch .float ,))
@@ -1175,7 +1263,19 @@ def test_vjpvmap(self, device, dtype, op):
1175
1263
xfail ('scatter_reduce' , 'mean' ),
1176
1264
xfail ('scatter_reduce' , 'prod' ),
1177
1265
skip ('linalg.householder_product' , '' , device_type = 'cuda' ), # flaky, I'm not sure why
1178
- xfail ('nn.functional.binary_cross_entropy_with_logits' )
1266
+ xfail ('nn.functional.binary_cross_entropy_with_logits' ),
1267
+ xfail ('__getitem__' , 'functorch' , device_type = 'cuda' ),
1268
+ xfail ('_masked.amax' , device_type = 'cuda' ),
1269
+ xfail ('_masked.amin' , device_type = 'cuda' ),
1270
+ xfail ('_masked.log_softmax' , device_type = 'cuda' ),
1271
+ xfail ('_masked.mean' , device_type = 'cuda' ),
1272
+ xfail ('_masked.norm' , device_type = 'cuda' ),
1273
+ xfail ('_masked.prod' , device_type = 'cuda' ),
1274
+ xfail ('_masked.softmax' , device_type = 'cuda' ),
1275
+ xfail ('_masked.softmin' , device_type = 'cuda' ),
1276
+ xfail ('_masked.std' , device_type = 'cuda' ),
1277
+ xfail ('_masked.sum' , device_type = 'cuda' ),
1278
+ xfail ('_masked.var' , device_type = 'cuda' ),
1179
1279
}))
1180
1280
def test_jvpvjp (self , device , dtype , op ):
1181
1281
if not op .supports_autograd :
0 commit comments