We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 67c0b89 commit bfe8b06Copy full SHA for bfe8b06
examples/models/llama/model.py
@@ -289,15 +289,6 @@ def __init__(self, **kwargs):
289
def get_eager_model(self) -> torch.nn.Module:
290
return self.model_
291
292
- if self.checkpoint_dtype:
293
- # convert to the type of the provided checkpoint
294
- # input and output are torch.long, so signature unchanged
295
- return self.model_.to(self.checkpoint_dtype)
296
- else:
297
- # int8 quantization code has some bf16,
298
- # switch all to FP32
299
- return self.model_.to(torch.float32)
300
-
301
def get_example_inputs(self):
302
if self.use_kv_cache:
303
return self.get_example_inputs_kvcache_sdpa()
0 commit comments