Questions about Transconv's attention design #7409
HelloWorldLTY
started this conversation in
General
Replies: 2 comments
-
I would suggest running the script with PyTorch profiler to see which part of the execution is taking so long: |
Beta Was this translation helpful? Give feedback.
0 replies
-
Also answered you on Slack. I think @akihironitta is right. From looking at the code, I am not sure your formula is correct. It looks like you will compute attention over all edges rather than local neighborhoods. I am also not sure |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi, I am trying to replace the transformconv's attention manchism with current fast attention design. However, the speed becomes slower, and my output is quite different.
The original code is:
My design is:
Beta Was this translation helpful? Give feedback.
All reactions