Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

FlashAttention2 implementation for OpenELM model #34485

Closed
2 tasks done
GorkaUrbizu opened this issue Oct 29, 2024 · 1 comment
Closed
2 tasks done

FlashAttention2 implementation for OpenELM model #34485

GorkaUrbizu opened this issue Oct 29, 2024 · 1 comment

Comments

@GorkaUrbizu
Copy link

Model description

OpenELM is already available on HF🤗 and transformers, but the model lacks support for flashAttention/FlashAttention2.

I'd love to have the flash attention available for OpenELM in the transformers enviroment.

thanks in advance.

Open source status

  • The model implementation is available
  • The model weights are available

Provide useful links for the implementation

No response

@Rocketknight1
Copy link
Member

Rocketknight1 commented Oct 29, 2024

Hi @GorkaUrbizu, OpenELM uses custom code, and so we can't actually add support in Transformers! However, I believe it already supports SDPA, which can dispatch to FlashAttention on newer versions of PyTorch.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants