-
-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Feature request: discussions around new features within HF ecosystem with unsloth #34
Comments
@younesbelkada Hey there! Saw many of ur PRs for HF - so great work again! So I actually saw SDPA support and I think I wrote a note in my benchmarks. For eg: Alpaca with SDPA on Tesla T4 ie:
I manually patched them for SDPA, and so on Tesla T4s, I did in fact benchmark SDPA (not native transformers, but just SDPA). Eg (The Flash Attention column is in fact SDPA)
So vs SDPA, Unsloth is 1.56x faster on a Tesla T4. |
But more than happy to collaborate on anything!! Again great work with TRL and PEFT! I'm actively following huggingface/transformers#26037 :) so that'll be massive for the next HF release! I'm also investigating LoftQ via PEFT as suggested by someone I was chatting with - I haven't tried it yet, but hopefully VRAM doesn't explode! |
Thanks very much for your positive reply @danielhanchen ! |
@younesbelkada Email is on my profile! :) |
Hi @danielhanchen
Thank you very much for this great project and pushing this forward for the community !
With TRL / PEFT team we've seen that your example scripts heavily rely on PEFT / TRL libraries and we wanted to see if you need any help or have any feature request around HF ecosystem we would be happy to collaborate and see what we can do together
Note also recently SDPA has been integrated into transformers core huggingface/transformers#26572 we were also wondering if you did some comparisons with unsloth against transformers 4.36.0
cc @pacman100 @lvwerra
The text was updated successfully, but these errors were encountered: