We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 6fabb91 commit e1371e4Copy full SHA for e1371e4
test/round.py
@@ -43,12 +43,13 @@ class TestFloatingPointRounding(unittest.TestCase):
43
("float8e4m3fn", __float8e4m3fn__, float8_e4m3fn, "cpu"),
44
("float16", __float16__, float16, "cpu"),
45
("bfloat16", __bfloat16__, bfloat16, "cpu"),
46
- ("float32", __float32__, float32, "cpu"),
+ # ("float32", __float32__, float32, "cpu"),
47
] + ([("float8e5m2", __float8e5m2__, float8_e5m2, "cuda"),
48
("float8e4m3fn", __float8e4m3fn__,float8_e4m3fn, "cuda"),
49
("float16", __float16__, float16, "cuda"),
50
("bfloat16", __bfloat16__, bfloat16, "cuda"),
51
- ("float32", __float32__, float32, "cuda")] if torch.cuda.is_available() else []))
+ # ("float32", __float32__, float32, "cuda"),
52
+ ] if torch.cuda.is_available() else []))
53
def test_rounding(self, name, fp, dtype, device):
54
a, b = finfo(dtype).min, finfo(dtype).max
55
assert a == fp.minimum and b == fp.maximum
0 commit comments