You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
[pt2e] Make prepare and convert faster by caching (#2983)
**Summary:** This is the torchao version of pytorch/pytorch#162550
by @navsud. Including the PR description here again:
D79674759 tried to fix the expensive prepare and convert steps,
as assert_and_get_unique_device was called multiple times.
This change fixes that issue by using functools.cache decorator.
**Test Plan:**
Verified on llm export to QNN.
LLM Quantization prepare time of ~20min reduced to ~3min.
0 commit comments