Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

epic: enabling GPU #1143

Open
5 tasks done
axel7083 opened this issue Jun 3, 2024 · 1 comment
Open
5 tasks done

epic: enabling GPU #1143

axel7083 opened this issue Jun 3, 2024 · 1 comment
Labels
area/inference kind/epic⚡ Issue describing an epic

Comments

@axel7083
Copy link
Contributor

axel7083 commented Jun 3, 2024

Epic domain

We need an issue to keep track of the requirements to support the GPUs in AI Lab

Detecting

Usage

Additional context

No response

@axel7083 axel7083 added kind/epic⚡ Issue describing an epic area/inference labels Jun 3, 2024
@axel7083
Copy link
Contributor Author

axel7083 commented Jun 3, 2024

List of issues encountered

Windows

ulimit problems

warning: failed to mlock 74489856-byte buffer (after previously locking 0 bytes): Cannot allocate memory
Try increasing RLIMIT_MEMLOCK ('ulimit -l' as root).

Solution: prefix export use_mlock=False before starting the llamacpp server

ref: abetlen/llama-cpp-python#708

Root requirement with the WSL

Need to run the container with --user 0 to let it access properly the GPU

@nichjones1 nichjones1 added this to the 1.3 milestone Aug 19, 2024
@nichjones1 nichjones1 moved this to 📋 Backlog in Podman Desktop Planning Aug 20, 2024
@deboer-tim deboer-tim removed this from the 1.3 milestone Oct 17, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
area/inference kind/epic⚡ Issue describing an epic
Projects
Status: 📋 Backlog
Development

No branches or pull requests

3 participants