Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Slight difference between implementation and paper. #6

Open
AI-Guru opened this issue Nov 6, 2019 · 2 comments
Open

Slight difference between implementation and paper. #6

AI-Guru opened this issue Nov 6, 2019 · 2 comments

Comments

@AI-Guru
Copy link

AI-Guru commented Nov 6, 2019

Namaste!

I think this is my final issue :D

In the implementation of the classifier, you start with a single-head-attention, then you do the spatial transformation followed by multi-head-attention. In the paper, on the other hand, you start with the transformation.

I have the feeling that I have missed something. What is your opinion?

@AI-Guru
Copy link
Author

AI-Guru commented Nov 6, 2019

Another difference. Correct me if I am wrong: The implementation of attn_feature does not 100% match. Especially "edge_feature = input_feature_tiled - neighbors" seems to be missing on the picture.

@AI-Guru
Copy link
Author

AI-Guru commented Nov 9, 2019

Yes. The paper mentions that a single-head attention happens as part of the spatial transform step.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant