-
Notifications
You must be signed in to change notification settings - Fork 487
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
longT5 BetterTransformer implementation #1506
Comments
seconding this! would be great |
Totally on board with this! Would love to see this feature added! |
@fxmarty can we try tackle this together? Thanks in advance |
Hi for reference we are upstreaming SDPA in Transformers, maybe it would be a better fit for longT5: huggingface/transformers#28005 Leaving this open as we may leverage nested tensors for longt5 (which are not in Transformers). |
Hi @ALL. Is this still open or you guys will work on it? |
Feature request
longT5 BetterTransformer implementation
Motivation
Encoder-decoder model trained on large context allow machine translation tasks
Your contribution
I looked at the implementation of regular T5 and it doesnt look to complex, i tried to implement myself but didnt succeed. If i can contribute please let me know.
Thank you,
Omri
The text was updated successfully, but these errors were encountered: