-
I have an AMD card rx580, I see that it NVIDA model but can I use AMD? |
Beta Was this translation helpful? Give feedback.
Replies: 3 comments 17 replies
-
It can't, unfortunately. The NVIDIA FasterTransformer library only supports NVIDIA GPUs. The Feature Roadmap includes supporting models from Huggingface Transformers, which I believe should enable AMD GPU support as well. |
Beta Was this translation helpful? Give feedback.
-
If you're remotely interested in ML, just buy a nvidia card. This will not be the only AI application that you will want to use and run into this problem with in the near future. |
Beta Was this translation helpful? Give feedback.
-
I was able to run fauxpilot on AMD GPU (RX 6700XT) by changing the Modified version: '3.3'
services:
triton:
build:
context: .
dockerfile: triton.Dockerfile
# HSA_OVERRIDE_GFX_VERSION=10.3.0 forces GFX version to 10.3.0 (this version appears to work with many AMD GPUs to run ROCm)
command: bash -c "HSA_OVERRIDE_GFX_VERSION=10.3.0 CUDA_VISIBLE_DEVICES=${GPUS} mpirun -n 1 --allow-run-as-root /opt/tritonserver/bin/tritonserver --model-repository=/model"
shm_size: '2gb'
volumes:
- ${MODEL_DIR}:/model
- ${HF_CACHE_DIR}:/root/.cache/huggingface
ports:
- "8000:8000"
- "${TRITON_PORT}:8001"
- "8002:8002"
# Since the NVIDIA driver passthrough doesn't work for AMD GPUs, we run the container in privileged mode (not optimal security-wise but working)
privileged: true
#deploy:
# resources:
# reservations:
# devices:
# - driver: nvidia
# count: all
# capabilities: [gpu]
copilot_proxy:
# For dockerhub version
# image: moyix/copilot_proxy:latest
# For local build
build:
context: .
dockerfile: proxy.Dockerfile
command: uvicorn app:app --host 0.0.0.0 --port 5000
env_file:
# Automatically created via ./setup.sh
- .env
ports:
- "${API_EXTERNAL_PORT}:5000" Modified FROM moyix/triton_with_ft:22.09
# Install dependencies: torch
# Here we use the latest version of pytorch with ROCm (see https://pytorch.org/get-started/locally/)
RUN python3 -m pip install --disable-pip-version-check -U torch --extra-index-url https://download.pytorch.org/whl/rocm5.4.2
RUN python3 -m pip install --disable-pip-version-check -U transformers bitsandbytes accelerate With only those modifications, you should be able to run fauxpilot on AMD GPUs. |
Beta Was this translation helpful? Give feedback.
It can't, unfortunately. The NVIDIA FasterTransformer library only supports NVIDIA GPUs. The Feature Roadmap includes supporting models from Huggingface Transformers, which I believe should enable AMD GPU support as well.