-
Notifications
You must be signed in to change notification settings - Fork 1k
This issue was moved to a discussion.
You can continue the conversation there. Go to discussion →
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Can't seem to get Tabby to run on GPU #2811
Comments
That i was get too, but only at builded tabby, when i am compiling 0.14.0 by my self it's works By the way, ROCm will work at RX6600.
|
Right! I never tried to build |
I realized I didn't compile for ROCm before, only for Vulkan. When I try to compile with
I use Gentoo and I should (as far as I know) have ROCm installed, but maybe I am missing something... I'll have to look into this more. (This happens both on |
A bit of a hacky solution that at least gets Tabby to compile: sudo mkdir /opt/rocm
sudo ln -sv /usr/lib/llvm/18 /opt/rocm/llvm However, besides seeming slower to generate responses in the chat, there's no difference. It still seems to run on the CPU when using: HSA_OVERRIDE_GFX_VERSION=10.3.0 cargo run --features rocm --release serve --model DeepseekCoder-1.3B --chat-model Qwen2-1.5B-Instruct --device rocm I also tried adding |
The fix for rocm is merged, the vulkan fix is probably similar (I haven't tested it): #2810 (comment) |
Does this fix also work on Windows? I'm running Tabby |
We are not distributing Windows binaries for ROCm at the moment, so it won't work. I recommend using the Vulkan backend for Windows if you have a non-Nvidia GPU card. |
This issue was moved to a discussion.
You can continue the conversation there. Go to discussion →
Describe the bug
(This might be me doing something wrong, and not a bug!)
I can't seem to get Tabby to run using my GPU (Radeon 6600 XT). Not with ROC, (which I believe is unsupported for my device) nor with Vulkan — which I believe should be supported?
Whenever I run tabby (
serve --model DeepseekCoder-1.3B --chat-model Qwen2-1.5B-Instruct --device vulkan
) and go into the Web UI to ask something I check my CPU and GPU usage (usingbtop
andamdgpu_top
respectively) I see my CPU usage spiking and almost no effect on the GPU. (This is when runningv0.15.0-r2
or compilingv0.15.0-r3
myself).If I try to use
v0.14.0
(same options) I instead get this:Information about your version
See above.
Information about your GPU
From
vulkaninfo
:From
rocminfo
(if it helps):Additional context
Full terminal output when starting Tabby (
v0.15.0-r3
, compiled myself):The text was updated successfully, but these errors were encountered: