Basic AMD ROCm setup guide #132
Replies: 5 comments 16 replies
-
Wow! After ZLUDA getting law suited to high heaven https://github.com/vosen/ZLUDA by Nvidia, I kind of thought getting ROCm to just work as a CUDA device would be dead in the water. You're not running ZLUDA, right? Just plain old ROCm (latest obviously) and the steps you said above? Either way is great news as I've not been able to test out any theories of getting an AMD card working (not having one personally). Feel free to feed back anything you notice and I have no idea if DeepSpeed will or wont work on an AMD flavour of CUDA, if the card is working with CUDA instructions, Ill give it a 30-40% chance it might, so you may be able to get a bit more of a speed up. Thanks so much for the instructions, hopefully someone else can give it a test at some point! :) If anyone else sees this and is willing to give it a test, please feed back and let us know how you get on! Thanks so much @shanecbauman! I really appreciate you taking the time to let me know how it worked/went and writing out some instructions. Really awesome. |
Beta Was this translation helpful? Give feedback.
-
Wow, ZLUDA looks great! Its a shame seeing compatibility efforts being stifled so much by the big players, hopefully progress on it will continue. But no im only using the latest ROCm libraries. I think the work AMD has done with HIP has really helped with making pytorch applications more portable between the GPU vendors. Ive had pretty good luck running automatic-1111s SD webui on my AMD card, so I was using the GPU prerequisites check in webui.sh as a reference for troubleshooting this. I will definitely be trying out deepspeed once I have more time to look into it, that speed improvement would really be nice. If I have any more success getting the rest of the features working Ill try to update here with my findings. |
Beta Was this translation helpful? Give feedback.
-
So I tried getting that working on a windows machine. Sadly ROCM ist not supported by windows, to be more clear PyTorch isn't. |
Beta Was this translation helpful? Give feedback.
-
Got that fixed, but cufft_internal_error is the next problem. Seems to be the problem that there's no PyTorch Version for ROCm for Windows. |
Beta Was this translation helpful? Give feedback.
-
Just to mention V1 Seems to work in Windows on WSL2 with Ubuntu 22.04 and Radeon 7900 XTX Uses my GPU Instructions: Got the requirements (already had it since I tried Stable Diffusion)
Then in alltalk-tts after its set up already
Then I also did this I think (based on SD Next instructions on Discord) "Patch PyTorch"
And then I needed to also copy this file into the lib folders overwriting the one there
just make sure to correct the path And start it up and then it works. This would take around a minute to generate on my 5950X CPU I also read some advices about using a newer kernel than the one in ubuntu 22.04 |
Beta Was this translation helpful? Give feedback.
-
First of all thanks for such a great program. The bulk TTS generator is exactly what I was looking for and I have been getting a ton of use out of it.
I have gotten this successfully setup with my RX 6750XT and I wanted to share my findings with anyone else who is thinking about running this on an AMD card. Ive only tested the regular model that is downloaded as part of the setup, and have not tried to implement deepspeed or done any model tuning as the vanilla model is sufficient for my use case:
./atsetup.sh
process, I installed the base requirements for AMD machines./start_alltalk.sh
and tts would work, but it was running on cpu and was very slow.cpu
orcuda
(GPU)cuda
.export HSA_OVERRIDE_GFX_VERSION=10.3.0
before launching alltalk in order to get it running properly. (I had to do this for my RX 6750XT)I saw that ROCm support is on the feature request list which is great to hear. Hopefully this information will help out anyone else who is running AMD.
ROCM 6.2 appears to have issues and downgrading to 6.0 is suggested (At this time). See this comment here
Beta Was this translation helpful? Give feedback.
All reactions