-
Notifications
You must be signed in to change notification settings - Fork 145
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
bug: models start the model imported can not work. #1439
Comments
Hi @cloudherder, for |
Thank you for your reply! You have created a great work! I tested it with an absolute path. The results are shown as follows: The following error is recorded in the Cortex.log file:
The sizes of the three models tested are 2.46GB,2.48GB and 7.06GB,my laptop has 16GB of memory, and using server.exe of llama.cpp can load and use these three models normally. |
@cloudherder Seems like you don't have enough VRAM. Please try to set the
Then set the
Run Can you share the output of |
|
Hi @cloudherder, apologies for late response. Can you please set |
Hi @cloudherder, we've released cortex v1.0.1 (release note) To update to cortex v1.0.1 (or download here: https://cortex.so/)
|
@cloudherder - closing this stale issue. We've released Cortex 1.0.3 with bugfixes an a much improved UX. |
Cortex version
cortex-1.0.0-rc1-windows-amd64-local-installer
Describe the Bug
Run all models by imported,return “ Model failed to load with status code: 500”
Steps to Reproduce
1.cortex-beta models import --model_id gemma-2b-Q8_0.gguf --model_path ./gemma-2b-Q8_0.gguf
It is successful, and can run modles subcommand ,such as list,get ,update, delete.
2.cortex-beta models start gemma-2b-Q8_0.gguf
It return :
gguf_init_from_file: failed to open '': 'Invalid argument'
{"timestamp":1728130117,"level":"ERROR","function":"LoadModel","line":186,"message":"llama.cpp unable to load model","model":""}
Model failed to load with status code: 500
Error: ?
Screenshots / Logs
cortex.log
cortex-cli.log
What is your OS?
What engine are you running?
The text was updated successfully, but these errors were encountered: