We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
We need an issue to keep track of the requirements to support the GPUs in AI Lab
LlamaCppPython
No response
The text was updated successfully, but these errors were encountered:
List of issues encountered
warning: failed to mlock 74489856-byte buffer (after previously locking 0 bytes): Cannot allocate memory Try increasing RLIMIT_MEMLOCK ('ulimit -l' as root).
Solution: prefix export use_mlock=False before starting the llamacpp server
export use_mlock=False
ref: abetlen/llama-cpp-python#708
Need to run the container with --user 0 to let it access properly the GPU
--user 0
Sorry, something went wrong.
No branches or pull requests
Epic domain
We need an issue to keep track of the requirements to support the GPUs in AI Lab
Detecting
Usage
LlamaCppPython
inference provider should support GPU on Windows WSL #1279Additional context
No response
The text was updated successfully, but these errors were encountered: