-
Notifications
You must be signed in to change notification settings - Fork 10.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Prompt structure after the --in-prefix-bos commit #2417
Comments
Previously, the first reverse prompt (if existed) is always inserted. Now if the model generate EOS, the reverse prompt is not inserted. To get the same behavior, please use
instead. |
Thank you for your response! I tested last night, and it appears to be mostly working. |
|
Sure, I appreciate the explanation. If I understand correctly then , I don't know how to dicern whether or not a model generates an EOS - most model cards don't contain that information, so I'd rather not guess and use an obselete parameter. |
It does exactly what it advertises to do, stopping generation when the code sees the string, which can be anything. If you work with models that do not generate EOS as needed (such as all the base models), you would need the reverse prompt. Of course the |
Okay, so previously More testing shows that whether or not In this case, I'm torn because newer llama.cpp has memory upgrades, but old llama.cpp works as expected, and so far I don't see a way to follow prompt structure strickly with many models. |
Please read Nous Hermes model card https://huggingface.co/NousResearch/Nous-Hermes-Llama2-13b It's an instruction model, not a chat model. After reading over the model card, you'll find a link to github that gives you example prompt. To strictly follow prompt structure, you unfortunately need to be able to read the Python code and reconstruct its prompt, https://github.com/teknium1/alpaca-roleplay-discordbot/blob/3595c171c2bab18feaa834518c1af990b978c49b/roleplay-bot.py#L116-L177 |
That's not the model, but I checked out the model card and it shows a Prompt Template section.
There was no need reconstruct a Prompt Template previously, and now decyphering Python is required? So @TheBloke's nous-hermes-llama-2-7b Prompt Template is incorrect? |
Actually mine might be incorrect in that I don't have a newline after instruction and they do. I'll fix that. Whatever is in the original model card will be correct, which is at the bottom of my readme. I'll update my PT section later this evening |
@TheBloke Thanks for clarifying. I tried with an updated PT and it works. @jxy I'll double check PT's going forward and close on this issue. Thank you. It's notable that the, |
Related: #2578 |
Hello, I'm adhereing to llama.cpp/example/main/README.txt
Here's main:
./main -m ~/wizardlm-7b-v1.0-uncensored.ggmlv3.q4_0.bin -i -r "User:" --in-prefix " " --in-suffix "Assistant:" -f ~/storage/shared/PT/Wiz.txt
Working as expected Grammar commit :
Note at the end of the Assistants message: llama.cpp inserts,
User:(space)
, compared with --in-prefix-bos commit:llama.cpp negates,
"User:"
, and only puts a space, breaking the structure of conversation.The directions in llama.cpp/example/main/README.txt do not work as expected since the --in-prefix-bos commit.
Thank you.
The text was updated successfully, but these errors were encountered: