diff --git a/examples/models/llama/export_llama_lib.py b/examples/models/llama/export_llama_lib.py index 4e004e773f2..89f8277bde7 100644 --- a/examples/models/llama/export_llama_lib.py +++ b/examples/models/llama/export_llama_lib.py @@ -800,6 +800,11 @@ def _export_llama(args) -> LLMEdgeManager: # noqa: C901 # pyre-fixme[16]: Module `backends` has no attribute `qualcomm`. canonicalize_program(builder.edge_manager.exported_program()) + print("lfq: exported program after to_backend, graph_module") + print(builder.edge_manager.exported_program().graph_module) + print("lfq: exported program after to_backend, print_readable") + print(builder.edge_manager.exported_program().graph_module.print_readable()) + builder = builder.to_executorch() if args.profile_memory: diff --git a/extension/llm/export/builder.py b/extension/llm/export/builder.py index ebc7f02ee1a..e67363d363e 100644 --- a/extension/llm/export/builder.py +++ b/extension/llm/export/builder.py @@ -193,6 +193,7 @@ def export(self) -> "LLMEdgeManager": dynamic_shapes=dynamic_shape, strict=True, ) + print("lfq: exported_module ", exported_module.graph_module) else: logging.info("Exporting with:") logging.info(f"inputs: {self.example_inputs}")