Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

bug: Model failed to load with status code: 500, could not load engine llamacpp #1422

Closed
dan-menlo opened this issue Oct 4, 2024 · 6 comments
Assignees
Milestone

Comments

@dan-menlo
Copy link
Contributor

dan-menlo commented Oct 4, 2024

Goal

  • A lot of useful error messages are currently not being bubbled up to CLI
  • We should bubble up as many error messages, to allow users to file more informative bugs or know what is going on

Current State

Running llama3.1 failed with model status code 500:

Image

The real error message is in the logs:

Image

@dan-menlo dan-menlo added this to Menlo Oct 4, 2024
@dan-menlo dan-menlo converted this from a draft issue Oct 4, 2024
@gabrielle-ong
Copy link
Contributor

gabrielle-ong commented Oct 4, 2024

v154 - also getting this error
image

Workaround

cortex engines install llama-cpp

which should have been pre-installed.

image

cc @hiento09 - discord discussion

@gabrielle-ong gabrielle-ong changed the title epic: cortex should bubble up better error messages bug: Model failed to load with status code: 500, could not load engine llamacpp Oct 5, 2024
@gabrielle-ong
Copy link
Contributor

gabrielle-ong commented Oct 5, 2024

should be solved by #1369?

@gabrielle-ong
Copy link
Contributor

Copying's Hien's post from #1396

I just tried a scenario as follows and encountered a similar issue to what you mentioned:
My machine is a Mac Silicon
I downloaded version 152, but the mac amd64 version – it still installed because my machine has Rosetta, but at this point, the llamacpp engine installed is the amd64 variant, which causes it to not run on Mac Silicon.
I ran cortex update to download version 154 universal and tried running cortex chat tinyllama:1b-gguf.

@gabrielle-ong gabrielle-ong moved this from Planning to Investigating in Menlo Oct 16, 2024
@gabrielle-ong
Copy link
Contributor

Proposal: Investigating -> Feature (high pain) -> Eng Planning

@gabrielle-ong
Copy link
Contributor

gabrielle-ong commented Oct 17, 2024

Decision from Dan / Nicole:

  • new Epic: bubble up error messages from Engines, highlight important error messages
  • Status: planning, Eng Spec: Planning, Sprints: 24/25

@gabrielle-ong
Copy link
Contributor

Closing this bug as resolved,
creating a separate discussion for bubbling error logs #1514

@github-project-automation github-project-automation bot moved this from Investigating to Review + QA in Menlo Oct 17, 2024
@gabrielle-ong gabrielle-ong moved this from Review + QA to Completed in Menlo Oct 17, 2024
@gabrielle-ong gabrielle-ong added this to the v1.0.1 milestone Oct 17, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
Archived in project
Development

No branches or pull requests

3 participants