Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

"failed to find free space in the KV cache" is displayed and no response is returned. #6603

Closed
Taikono-Himazin opened this issue Apr 11, 2024 · 1 comment

Comments

@Taikono-Himazin
Copy link
Contributor

Taikono-Himazin commented Apr 11, 2024

When using llama.cpp via LocalAI, "failed to find free space in the KV cache" will be displayed after using it for a while, and the string that can respond will gradually become shorter, and eventually it will not be possible to respond.

I looked at past issues such as #4185, but I don't really understand how to solve them.

I would like to delete the KV cache or increase the capacity of the KV cache, but how should I do that?

The version of llama.cpp seems to be 8228b66.

@phymbert
Copy link
Collaborator

Please upgrade the KV cache size yes using --ctx-size

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants