-
Notifications
You must be signed in to change notification settings - Fork 747
Improve softmax perf when transpose is not needed #13081
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/13081
Note: Links to docs will display an error until the docs builds have been completed. ❌ 1 New FailureAs of commit ceb8d36 with merge base fe84495 ( NEW FAILURE - The following job has failed:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
|
This pull request was exported from Phabricator. Differential Revision: D79514231 |
This PR needs a
|
e9839dd to
c96d19d
Compare
|
This pull request was exported from Phabricator. Differential Revision: D79514231 |
Summary: When the supplied dimension is the last dim of the tensor, we don't need to permute anything and can call the nnlib kernel directly. Differential Revision: D79514231
c96d19d to
1ab590c
Compare
|
This pull request was exported from Phabricator. Differential Revision: D79514231 |
Summary: Pull Request resolved: #13081 When the supplied dimension is the last dim of the tensor, we don't need to permute anything and can call the nnlib kernel directly. Differential Revision: D79514231
1ab590c to
3e0bf0c
Compare
Summary: When the supplied dimension is the last dim of the tensor, we don't need to permute anything and can call the nnlib kernel directly. Differential Revision: D79514231
3e0bf0c to
c39cede
Compare
Summary: Pull Request resolved: #13081 When the supplied dimension is the last dim of the tensor, we don't need to permute anything and can call the nnlib kernel directly. Differential Revision: D79514231
|
This pull request was exported from Phabricator. Differential Revision: D79514231 |
c39cede to
ceb8d36
Compare
Differential Revision: D79514231 Pull Request resolved: pytorch#13081
Differential Revision: D79514231