Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Does mistral support multiple gpus? #451

Closed
NiuBlibing opened this issue Jun 20, 2024 · 3 comments
Closed

Does mistral support multiple gpus? #451

NiuBlibing opened this issue Jun 20, 2024 · 3 comments
Labels
bug Something isn't working

Comments

@NiuBlibing
Copy link

Describe the bug
I build mistral with cargo build --release --features "cuda flash-attn" and run model with ./target/release/mistralrs-server --port 1234 -n 8 plain -m ./Qwen/Qwen2-72B-Instruct/ -a qwen2 on 8*a100 device, the nvitop shown only one gpu' memory is growing and then oom.

Latest commit
3a79137

@NiuBlibing NiuBlibing added the bug Something isn't working label Jun 20, 2024
@b0xtch
Copy link

b0xtch commented Jun 21, 2024

#375

I don't think so

@NiuBlibing
Copy link
Author

#395

@EricLBuehler
Copy link
Owner

Hi @NiuBlibing! Multiple GPUs is not supported yet, but I will add cross-GPU mapping support first before NCCL support.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

3 participants