Skip to content

Commit f162b53

Browse files
committed
updates
1 parent ebf1db5 commit f162b53

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

tests/models/test_modeling_common.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1416,8 +1416,8 @@ def get_memory_usage(storage_dtype, compute_dtype):
14161416

14171417
compute_capability = get_torch_cuda_device_capability()
14181418
self.assertTrue(fp8_e4m3_bf16_memory_footprint < fp8_e4m3_fp32_memory_footprint < fp32_memory_footprint)
1419-
# NOTE: the following assertion will fail on our CI (running Tesla T4) due to bf16 using more memory than fp32.
1420-
# On other devices, such as DGX (Ampere) and Audace (Ada), the test passes.
1419+
# NOTE: the following assertion would fail on our CI (running Tesla T4) due to bf16 using more memory than fp32.
1420+
# On other devices, such as DGX (Ampere) and Audace (Ada), the test passes. So, we conditionally check it.
14211421
if compute_capability and compute_capability >= LEAST_COMPUTE_CAPABILITY:
14221422
self.assertTrue(fp8_e4m3_bf16_max_memory < fp8_e4m3_fp32_max_memory)
14231423
# On this dummy test case with a small model, sometimes fp8_e4m3_fp32 max memory usage is higher than fp32 by a few

0 commit comments

Comments
 (0)