-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
inference problem with baichuan 13b #9894
Comments
The current machine being used is an a770, and the GPU memory should be sufficient. I hope you can provide me with some guidance. |
Could you provide more details?
|
I am using baichuan2 and the sequence length should be the default. |
https://github.com/intel-analytics/BigDL/blob/main/python/llm/example/GPU/HF-Transformers-AutoModels/Model/baichuan2/generate.py#L50 |
Which script are you using and what is the default? Is it this one? https://github.com/intel-analytics/BigDL/tree/main/python/llm/dev/benchmark/all-in-one |
I just download the baichuan2-13b model from HF and run model.chat. This is what I mean default |
Does |
|
I try to use xpu:0 and xpu:1 teo different situation. In xpu:1 there will come a problem that the device_id is out of range, and xpu:0 is the original state. What can I do next. |
So are there multiple gpu cards on your machine? After sourcing oneapi, you can use sycl-ls to check the gpu cards on your machine: |
|
Seems only one GPU is detected... Are other gpus properly set? |
Not sure why there only one gpu detected, I see the gpu 2 in this figure? |
It looks like your driver(released in 2023.7) is a little old. Please update your driver to latest version and try again. |
Hi, I have tested it on my side using bigdl and model.chat from HF. And it worked fine. But I am a bit curious about the log output
|
not sure what is going wrong here
The text was updated successfully, but these errors were encountered: