Formula of Graph Attention Networks with Edge Features #8846
-
Hi, I'm using Graph Attention Networks (GAT), and I saw the implementation here. The calculation of attention coefficients is done by summing up parts of the node representations instead of concatenation. I read that this helps to avoid the memory consumption issue.. My question is: how do you write the math formula that involves the concatenation of edge features? From the implementation code, it seems that the weight matrix and attention vector for the edge feature has different shape compared to the two used for node representations. Original: Should I write: Thank you. |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment
-
It would be
|
Beta Was this translation helpful? Give feedback.
It would be