You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi @GorkaUrbizu, OpenELM uses custom code, and so we can't actually add support in Transformers! However, I believe it already supports SDPA, which can dispatch to FlashAttention on newer versions of PyTorch.
Model description
OpenELM is already available on HF🤗 and transformers, but the model lacks support for flashAttention/FlashAttention2.
I'd love to have the flash attention available for OpenELM in the transformers enviroment.
thanks in advance.
Open source status
Provide useful links for the implementation
No response
The text was updated successfully, but these errors were encountered: