Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Q LORA or LORA implementation #50

Open
okpatil4u opened this issue Jul 4, 2023 · 4 comments
Open

Q LORA or LORA implementation #50

okpatil4u opened this issue Jul 4, 2023 · 4 comments

Comments

@okpatil4u
Copy link

Feature description

For finetuning existing text generation model, LORA and QLORA are popularly being used. Can we create pipelines for download models from Huggingface, then finetune the models using LORA or QLORA ?

Feature motivation

For text generation models, finetuning is preferred, rather than training from scratch. LORA or QLORA reduces the time required for finetuning by an order of magnitude.

@antimora
Copy link
Collaborator

antimora commented Jul 6, 2023

We plan to host another repository to automate importing models. Currently we are working on burn-import to support more ONNX OPs. We plan to start with a few popular examples and gradually move to other pre-trained models.

The idea is that once you import your model, you can, in addition to inference, finetune or do additional complex training. That's the whole idea for burn-import.

At the beginning, we will support ONNX and other data formats, such as pth file format.

If you wish to contribute in such effort, we will happily assist you. burn-import has been cleaned up and ready for additional contribution.

@nathanielsimard if you have anything to add, please let us know.

@airstrike
Copy link

Hi, just curious if any news since the last update was a while ago. Thank you!

@antimora antimora transferred this issue from tracel-ai/burn Nov 25, 2024
@antimora
Copy link
Collaborator

CCing @laggui

@laggui
Copy link
Member

laggui commented Nov 26, 2024

The plan was to add those after Llama was added, but some plans changed since we introduced CubeCL 🙂

I think it's safe to say that there will be a positive update sometime in ~3 months regarding this 👀 Stay tuned.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants