Skip to content

Commit

Permalink
Merge pull request #704 from axel7083/fix/cuda-build
Browse files Browse the repository at this point in the history
fix(modell_server): deprecated make arguments for llamacpp server
  • Loading branch information
rhatdan authored Jul 30, 2024
2 parents e273eae + f60606d commit 7409fdc
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion model_servers/llamacpp_python/cuda/Containerfile
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ RUN dnf install -y gcc-toolset-13-gcc gcc-toolset-13-gcc-c++
USER 1001
WORKDIR /locallm
COPY src .
ENV CMAKE_ARGS="-DLLAMA_CUBLAS=on -DLLAMA_AVX2=OFF -DLLAMA_FMA=OFF -DLLAMA_F16C=OFF"
ENV CMAKE_ARGS="-DGGML_CUDA=on -DLLAMA_AVX2=OFF -DLLAMA_FMA=OFF -DLLAMA_F16C=OFF"
ENV FORCE_CMAKE=1
RUN CC="/opt/rh/gcc-toolset-13/root/usr/bin/gcc" CXX="/opt/rh/gcc-toolset-13/root/usr/bin/g++" pip install --no-cache-dir -r ./requirements.txt
ENTRYPOINT [ "sh", "run.sh" ]

0 comments on commit 7409fdc

Please sign in to comment.