Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: add yayi2-30b example #208

Closed
wants to merge 3 commits into from
Closed

feat: add yayi2-30b example #208

wants to merge 3 commits into from

Conversation

mzbac
Copy link
Contributor

@mzbac mzbac commented Dec 31, 2023

Yayi2-30b has a score of over 80 MMLU. I performed some fine-tuning on it using qlora, and from my quick test, it appears to be very promising. so, I created the mlx example for this model in case anyone wants to run it via mlx. However, this model has an unusual k,v layer that causes the quantization to fail. Currently, I haven't found any quantization tools that support this model (except for bnb nf4). It would be great if mlx could provide support for its quantization.

FYI:
https://huggingface.co/wenge-research/yayi2-30b
https://huggingface.co/mzbac/yayi2-30b-guanaco
ml-explore/mlx#328

@mzbac
Copy link
Contributor Author

mzbac commented Jan 1, 2024

I have added the workaround for the quant as suggested in ml-explore/mlx#328. Now, the example works with 4-bit quantization. Once the PR gets merged, I will upload the 4-bit quantized model.

demo.mov

@awni
Copy link
Member

awni commented Jan 10, 2024

HI @mzbac sorry for the delayed review here. Do you still want to merge this? I think given the non-standard size it wouldn't fit easily in our hf_llm example, but wdyt?

@mzbac
Copy link
Contributor Author

mzbac commented Jan 10, 2024

I think this should be supported by hf_llm once we fix the quant non-32 dimension. I'm happy to close this one. Meanwhile, if people want to try this model in f16 precision, they should be able to run it via hf_llm.

@awni
Copy link
Member

awni commented Jan 10, 2024

Sounds good, thank you!

@awni awni closed this Jan 10, 2024
@mzbac mzbac deleted the yayi branch January 13, 2024 05:07
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants