-
Notifications
You must be signed in to change notification settings - Fork 285
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Draft: feat: Support Llama 3 model (#478) #479
Draft: feat: Support Llama 3 model (#478) #479
Conversation
0ec7c86
to
e91a278
Compare
aa92aab
to
ecb270e
Compare
Is there any more work to be expected here, or can we perhaps remove the "Draft" prefix? |
As far as I can see it's finished (can you double-check the InfillPrompt?), I'm just waiting for the llama.cpp server support 😅 |
Nice! It looks like the model doesn't support infilling, or at least I couldn't find anything. Maybe we can just remove the new infill template for now. |
Really? All |
Hmm, their model card doesn't specify anything related to that matter - https://llama.meta.com/docs/model-cards-and-prompt-formats/meta-llama-3 |
All I can see is that there is a lot of confusion about the model, some people have many problems, some have none... I guess it depends on how you use it, and some HF models also seem to contain problems. |
No, but we were using it for a while, I think. However, I can't remember the reason why we switched back. |
* feat: Support Llama 3 model (#478) * Use new InfillPrompt * Switch to lmstudio-community * Use new Prompt * llama.cpp removed the BOS token ggml-org/llama.cpp@a55d8a9 * Add tests * I would prefer a stream based solution * Add 70B models * Add tests for skipping blank system prompt * Remove InfillPrompt for now
Wow, fresh of the press, seems to be one of the best models now!! 🚀
Waiting for llama.cpp support... 😅
ggml-org/llama.cpp#6747
ggml-org/llama.cpp#6751