You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In the implementation of the classifier, you start with a single-head-attention, then you do the spatial transformation followed by multi-head-attention. In the paper, on the other hand, you start with the transformation.
I have the feeling that I have missed something. What is your opinion?
The text was updated successfully, but these errors were encountered:
Another difference. Correct me if I am wrong: The implementation of attn_feature does not 100% match. Especially "edge_feature = input_feature_tiled - neighbors" seems to be missing on the picture.
Namaste!
I think this is my final issue :D
In the implementation of the classifier, you start with a single-head-attention, then you do the spatial transformation followed by multi-head-attention. In the paper, on the other hand, you start with the transformation.
I have the feeling that I have missed something. What is your opinion?
The text was updated successfully, but these errors were encountered: