We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent d4b88d7 commit 30f13edCopy full SHA for 30f13ed
tests/quantization/gguf/test_gguf.py
@@ -82,6 +82,7 @@ def test_gguf_memory(self):
82
self.ckpt_path, quantization_config=quantization_config, torch_dtype=self.torch_dtype
83
)
84
model.to("cuda")
85
+ assert (model.get_memory_footprint() / 1024**3) < 5
86
inputs = self.get_dummy_inputs()
87
88
torch.cuda.reset_peak_memory_stats()
0 commit comments