-
Notifications
You must be signed in to change notification settings - Fork 688
Use a symmetric quantization with no clipping error to improve llama perplexity #5163
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/5163
Note: Links to docs will display an error until the docs builds have been completed. ✅ No FailuresAs of commit 04e2465 with merge base a4be79f ( This comment was automatically generated by Dr. CI and updates every 15 minutes. |
@iseeyuan has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
…perplexity (#5163) Summary: Refer to pytorch/ao#805 for the details. With this change, the perplexity of a llama model is improved 4% on wikitext. Differential Revision: D62342523 Pulled By: iseeyuan
6785322
to
6de5818
Compare
This pull request was exported from Phabricator. Differential Revision: D62342523 |
…perplexity (#5163) Summary: Refer to pytorch/ao#805 for the details. With this change, the perplexity of a llama model is improved 4% on wikitext. Differential Revision: D62342523 Pulled By: iseeyuan
6de5818
to
129d673
Compare
This pull request was exported from Phabricator. Differential Revision: D62342523 |
…perplexity (#5163) Summary: Refer to pytorch/ao#805 for the details. With this change, the perplexity of a llama model is improved 4% on wikitext. Reviewed By: mergennachin, helunwencser Differential Revision: D62342523 Pulled By: iseeyuan
129d673
to
2db6f07
Compare
This pull request was exported from Phabricator. Differential Revision: D62342523 |
…perplexity (#5163) Summary: Refer to pytorch/ao#805 for the details. With this change, the perplexity of a llama model is improved 4% on wikitext. Reviewed By: mergennachin, helunwencser Differential Revision: D62342523 Pulled By: iseeyuan
2db6f07
to
04e2465
Compare
This pull request was exported from Phabricator. Differential Revision: D62342523 |
Looks like this PR hasn't been updated in a while so we're going to go ahead and mark this as |
Refer to pytorch/ao#805 for the details.
With this change, the perplexity of a llama model is improved 4% on wikitext.