-
Notifications
You must be signed in to change notification settings - Fork 10.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
unexpected shut down when number of tokens is large #134
Comments
I have the same problem as you, and here is my tests I have changed the parameter
256 Limit (Almost 8 lines)
1048 Limit (Almost 12 lines)
|
Duplicate of #71
Originally posted by @eous in #71 (comment) |
problem solved thanks to @AndrewKeYanzhe's help, here is solution: in the file in to (number size can be adjusted according to RAM):
the code should be able to give a 2048-words of context reference then run the code in terminal:
to re-compile |
add llama2.c-android to readme
I found that the model of LLaMA-7B shut down unexpectedly when the number of tokens in prompt reaches some value, this value is approximately to be 500
this cannot be solved by setting number of tokens to predict high (e.g. 204800)
my initialization is:
where
p.txt
is a file containing some prompts, and the token number of prompts ismain: number of tokens in prompt = 486
the program shut down unexpectedly after a few interactions, last shows:
I am using macPro M1 with 16GB RAM
I am wondering is there any limitation in the program or did i do something wrong
The text was updated successfully, but these errors were encountered: