Skip to content
New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

efficient implementation #5

Open
amrhamedp opened this issue Jan 30, 2022 · 2 comments
Open

efficient implementation #5

amrhamedp opened this issue Jan 30, 2022 · 2 comments

Comments

@amrhamedp
Copy link

Hi,
I wonder if relative distances and coordinates can be handled more efficiently using memory efficient attention as in " Self-attention Does Not Need O(n^2) Memory". It is straightforward for the scalar part.

@lucidrains
Copy link
Owner

yea definitely

have you found En-transformer useful?

@amrhamedp
Copy link
Author

yes ...it is pretty useful however I believe that distances should be used to allow nodes positional encoding. I mean something like IPA (invariant point attention) or schnet like encoding for distances using RBF.

# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants