Skip to content

Commit a85ce2a

Browse files
committed
Remove test stuff
1 parent 2ced59e commit a85ce2a

File tree

1 file changed

+0
-8
lines changed

1 file changed

+0
-8
lines changed

examples/models/llama/export_llama_lib.py

Lines changed: 0 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -619,14 +619,6 @@ def _prepare_for_llama_export(args) -> LLMEdgeManager:
619619
)
620620
)
621621

622-
quantized = torch.load("/home/jackzhxng/torchrepos/executorch/fake_quantized_weights.pt")
623-
# Test run the model to trace.
624-
edge_manager.model(
625-
torch.tensor([[2, 3, 4]], dtype=torch.long),
626-
{"input_pos": torch.tensor([0], dtype=torch.long)},
627-
)
628-
# torch.testing.assert_close()
629-
630622
# We want to do compute the actual ops in the precision of the dtype_override.
631623
def _set_precision_to_fp32(module):
632624
"""

0 commit comments

Comments
 (0)