Skip to content

Commit

Permalink
fix torch requirements
Browse files Browse the repository at this point in the history
  • Loading branch information
Josh-XT committed Jul 30, 2024
1 parent 7428a54 commit 679ee9d
Show file tree
Hide file tree
Showing 2 changed files with 4 additions and 5 deletions.
2 changes: 0 additions & 2 deletions cuda-requirements.txt
Original file line number Diff line number Diff line change
Expand Up @@ -9,8 +9,6 @@ beautifulsoup4==4.12.3
faster-whisper==1.0.3
pydub==0.25.1
ffmpeg-python==0.2.0
torch==2.3.1+cu121
torchaudio==2.3.1+cu121
transformers
TTS==0.22.0
sounddevice==0.4.6
Expand Down
7 changes: 4 additions & 3 deletions cuda.Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -13,7 +13,9 @@ ENV HOST=0.0.0.0 \
LLAMA_CUBLAS=1
COPY cuda-requirements.txt .
RUN python3 -m pip install --upgrade pip cmake scikit-build setuptools wheel --no-cache-dir && \
CMAKE_ARGS="-DLLAMA_CUDA=on" FORCE_CMAKE=1 pip install llama-cpp-python==0.2.83 --no-cache-dir && \
pip install https://download.pytorch.org/whl/nightly/cu121/torch-2.2.0.dev20231010%2Bcu121-cp310-cp310-linux_x86_64.whl && \
pip install https://download.pytorch.org/whl/nightly/cu121/torchaudio-2.2.0.dev20231010%2Bcu121-cp310-cp310-linux_x86_64.whl && \
pip install llama-cpp-python --extra-index-url https://abetlen.github.io/llama-cpp-python/whl/cu123 && \
pip install --no-cache-dir -r cuda-requirements.txt
RUN git clone https://github.com/Josh-XT/DeepSeek-VL deepseek && \
cd deepseek && \
Expand All @@ -23,5 +25,4 @@ RUN pip install spacy==3.7.4 && \
python -m spacy download en_core_web_sm
COPY . .
EXPOSE 8091
EXPOSE 8502
CMD streamlit run ui.py & uvicorn app:app --host 0.0.0.0 --port 8091 --workers 1 --proxy-headers
CMD uvicorn app:app --host 0.0.0.0 --port 8091 --workers 1 --proxy-headers

0 comments on commit 679ee9d

Please sign in to comment.