diff --git a/AudioQnA/docker/gaudi/README.md b/AudioQnA/docker/gaudi/README.md index 999a1ac4eb..6c3eb0b6f4 100644 --- a/AudioQnA/docker/gaudi/README.md +++ b/AudioQnA/docker/gaudi/README.md @@ -14,24 +14,24 @@ cd GenAIComps ### 2. Build ASR Image ```bash -docker build -t opea/whisper-gaudi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/asr/whisper/Dockerfile_hpu . +docker build -t opea/whisper-gaudi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/asr/whisper/Dockerfile_hpu . -docker build -t opea/asr:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/asr/Dockerfile . +docker build -t opea/asr:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/asr/Dockerfile . ``` ### 3. Build LLM Image ```bash -docker build --no-cache -t opea/llm-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . +docker build --no-cache -t opea/llm-tgi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . ``` ### 4. Build TTS Image ```bash -docker build -t opea/speecht5-gaudi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/tts/speecht5/Dockerfile_hpu . +docker build -t opea/speecht5-gaudi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/tts/speecht5/Dockerfile_hpu . -docker build -t opea/tts:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/tts/Dockerfile . +docker build -t opea/tts:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/tts/Dockerfile . ``` ### 6. Build MegaService Docker Image @@ -41,17 +41,17 @@ To construct the Mega Service, we utilize the [GenAIComps](https://github.com/op ```bash git clone https://github.com/opea-project/GenAIExamples.git cd GenAIExamples/AudioQnA/docker -docker build --no-cache -t opea/audioqna:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . +docker build --no-cache -t opea/audioqna:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . ``` Then run the command `docker images`, you will have following images ready: -1. `opea/whisper-gaudi:latest` -2. `opea/asr:latest` -3. `opea/llm-tgi:latest` -4. `opea/speecht5-gaudi:latest` -5. `opea/tts:latest` -6. `opea/audioqna:latest` +1. `opea/whisper-gaudi:latest.1.0` +2. `opea/asr:latest.1.0` +3. `opea/llm-tgi:latest.1.0` +4. `opea/speecht5-gaudi:latest.1.0` +5. `opea/tts:latest.1.0` +6. `opea/audioqna:latest.1.0` ## πŸš€ Set the environment variables @@ -81,7 +81,7 @@ export LLM_SERVICE_PORT=3007 ```bash cd GenAIExamples/AudioQnA/docker/gaudi/ -docker compose up -d +TAG=0.1.0 docker compose up -d ``` ## πŸš€ Test MicroServices diff --git a/AudioQnA/docker/xeon/README.md b/AudioQnA/docker/xeon/README.md index 3537906e14..70eb23d3f7 100644 --- a/AudioQnA/docker/xeon/README.md +++ b/AudioQnA/docker/xeon/README.md @@ -14,24 +14,24 @@ cd GenAIComps ### 2. Build ASR Image ```bash -docker build -t opea/whisper:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/asr/whisper/Dockerfile . +docker build -t opea/whisper:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/asr/whisper/Dockerfile . -docker build -t opea/asr:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/asr/Dockerfile . +docker build -t opea/asr:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/asr/Dockerfile . ``` ### 3. Build LLM Image ```bash -docker build --no-cache -t opea/llm-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . +docker build --no-cache -t opea/llm-tgi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . ``` ### 4. Build TTS Image ```bash -docker build -t opea/speecht5:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/tts/speecht5/Dockerfile . +docker build -t opea/speecht5:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/tts/speecht5/Dockerfile . -docker build -t opea/tts:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/tts/Dockerfile . +docker build -t opea/tts:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/tts/Dockerfile . ``` ### 6. Build MegaService Docker Image @@ -41,17 +41,17 @@ To construct the Mega Service, we utilize the [GenAIComps](https://github.com/op ```bash git clone https://github.com/opea-project/GenAIExamples.git cd GenAIExamples/AudioQnA/docker -docker build --no-cache -t opea/audioqna:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . +docker build --no-cache -t opea/audioqna:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . ``` Then run the command `docker images`, you will have following images ready: -1. `opea/whisper:latest` -2. `opea/asr:latest` -3. `opea/llm-tgi:latest` -4. `opea/speecht5:latest` -5. `opea/tts:latest` -6. `opea/audioqna:latest` +1. `opea/whisper:latest.1.0` +2. `opea/asr:latest.1.0` +3. `opea/llm-tgi:latest.1.0` +4. `opea/speecht5:latest.1.0` +5. `opea/tts:latest.1.0` +6. `opea/audioqna:latest.1.0` ## πŸš€ Set the environment variables @@ -81,7 +81,7 @@ export LLM_SERVICE_PORT=3007 ```bash cd GenAIExamples/AudioQnA/docker/xeon/ -docker compose up -d +TAG=0.1.0 docker compose up -d ``` ## πŸš€ Test MicroServices diff --git a/AudioQnA/kubernetes/README.md b/AudioQnA/kubernetes/README.md index 569ea9a7ad..7fae4dd969 100644 --- a/AudioQnA/kubernetes/README.md +++ b/AudioQnA/kubernetes/README.md @@ -15,19 +15,19 @@ The AudioQnA application is defined as a Custom Resource (CR) file that the abov The AudioQnA uses the below prebuilt images if you choose a Xeon deployment - tgi-service: ghcr.io/huggingface/text-generation-inference:1.4 -- llm: opea/llm-tgi:latest -- asr: opea/asr:latest -- whisper: opea/whisper:latest -- tts: opea/tts:latest -- speecht5: opea/speecht5:latest +- llm: opea/llm-tgi:latest.1.0 +- asr: opea/asr:latest.1.0 +- whisper: opea/whisper:latest.1.0 +- tts: opea/tts:latest.1.0 +- speecht5: opea/speecht5:latest.1.0 Should you desire to use the Gaudi accelerator, two alternate images are used for the embedding and llm services. For Gaudi: - tgi-service: ghcr.io/huggingface/tgi-gaudi:1.2.1 -- whisper-gaudi: opea/whisper-gaudi:latest -- speecht5-gaudi: opea/speecht5-gaudi:latest +- whisper-gaudi: opea/whisper-gaudi:latest.1.0 +- speecht5-gaudi: opea/speecht5-gaudi:latest.1.0 > [NOTE] > Please refer to [Xeon README](https://github.com/opea-project/GenAIExamples/blob/main/AudioQnA/docker/xeon/README.md) or [Gaudi README](https://github.com/opea-project/GenAIExamples/blob/main/AudioQnA/docker/gaudi/README.md) to build the OPEA images. These too will be available on Docker Hub soon to simplify use. diff --git a/AudioQnA/kubernetes/manifests/gaudi/audioqna.yaml b/AudioQnA/kubernetes/manifests/gaudi/audioqna.yaml index 5a824ef3e6..c5c4c243e8 100644 --- a/AudioQnA/kubernetes/manifests/gaudi/audioqna.yaml +++ b/AudioQnA/kubernetes/manifests/gaudi/audioqna.yaml @@ -50,7 +50,7 @@ spec: - envFrom: - configMapRef: name: audio-qna-config - image: opea/asr:latest + image: opea/asr:latest.1.0 imagePullPolicy: IfNotPresent name: asr-deploy args: null @@ -101,7 +101,7 @@ spec: - envFrom: - configMapRef: name: audio-qna-config - image: opea/whisper-gaudi:latest + image: opea/whisper-gaudi:latest.1.0 imagePullPolicy: IfNotPresent name: whisper-deploy args: null @@ -164,7 +164,7 @@ spec: - envFrom: - configMapRef: name: audio-qna-config - image: opea/tts:latest + image: opea/tts:latest.1.0 imagePullPolicy: IfNotPresent name: tts-deploy args: null @@ -215,7 +215,7 @@ spec: - envFrom: - configMapRef: name: audio-qna-config - image: opea/speecht5-gaudi:latest + image: opea/speecht5-gaudi:latest.1.0 imagePullPolicy: IfNotPresent name: speecht5-deploy args: null @@ -365,7 +365,7 @@ spec: - envFrom: - configMapRef: name: audio-qna-config - image: opea/llm-tgi:latest + image: opea/llm-tgi:latest.1.0 imagePullPolicy: IfNotPresent name: llm-deploy args: null @@ -416,7 +416,7 @@ spec: - envFrom: - configMapRef: name: audio-qna-config - image: opea/audioqna:latest + image: opea/audioqna:latest.1.0 imagePullPolicy: IfNotPresent name: audioqna-backend-server-deploy args: null diff --git a/AudioQnA/kubernetes/manifests/xeon/audioqna.yaml b/AudioQnA/kubernetes/manifests/xeon/audioqna.yaml index 89b3e6b1a4..a908d8f361 100644 --- a/AudioQnA/kubernetes/manifests/xeon/audioqna.yaml +++ b/AudioQnA/kubernetes/manifests/xeon/audioqna.yaml @@ -50,7 +50,7 @@ spec: - envFrom: - configMapRef: name: audio-qna-config - image: opea/asr:latest + image: opea/asr:latest.1.0 imagePullPolicy: IfNotPresent name: asr-deploy args: null @@ -101,7 +101,7 @@ spec: - envFrom: - configMapRef: name: audio-qna-config - image: opea/whisper:latest + image: opea/whisper:latest.1.0 imagePullPolicy: IfNotPresent name: whisper-deploy args: null @@ -152,7 +152,7 @@ spec: - envFrom: - configMapRef: name: audio-qna-config - image: opea/tts:latest + image: opea/tts:latest.1.0 imagePullPolicy: IfNotPresent name: tts-deploy args: null @@ -203,7 +203,7 @@ spec: - envFrom: - configMapRef: name: audio-qna-config - image: opea/speecht5:latest + image: opea/speecht5:latest.1.0 imagePullPolicy: IfNotPresent name: speecht5-deploy args: null @@ -321,7 +321,7 @@ spec: - envFrom: - configMapRef: name: audio-qna-config - image: opea/llm-tgi:latest + image: opea/llm-tgi:latest.1.0 imagePullPolicy: IfNotPresent name: llm-deploy args: null @@ -372,7 +372,7 @@ spec: - envFrom: - configMapRef: name: audio-qna-config - image: opea/audioqna:latest + image: opea/audioqna:latest.1.0 imagePullPolicy: IfNotPresent name: audioqna-backend-server-deploy args: null diff --git a/ChatQnA/README.md b/ChatQnA/README.md index 43579d0968..a4d070c7e7 100644 --- a/ChatQnA/README.md +++ b/ChatQnA/README.md @@ -23,7 +23,7 @@ Currently we support two ways of deploying ChatQnA services with docker compose: 1. Start services using the docker image on `docker hub`: ```bash - docker pull opea/chatqna:latest + docker pull opea/chatqna:latest.1.0 ``` Two type of UI are supported now, choose one you like and pull the referred docker image. @@ -31,9 +31,9 @@ Currently we support two ways of deploying ChatQnA services with docker compose: If you choose conversational UI, follow the [instruction](https://github.com/opea-project/GenAIExamples/tree/main/ChatQnA/docker/gaudi#-launch-the-conversational-ui-optional) and modify the [compose.yaml](./docker/xeon/compose.yaml). ```bash - docker pull opea/chatqna-ui:latest + docker pull opea/chatqna-ui:latest.1.0 # or - docker pull opea/chatqna-conversation-ui:latest + docker pull opea/chatqna-conversation-ui:latest.1.0 ``` 2. Start services using the docker images `built from source`: [Guide](./docker) @@ -78,7 +78,7 @@ Find the corresponding [compose.yaml](./docker/gaudi/compose.yaml). ```bash cd GenAIExamples/ChatQnA/docker/gaudi/ -docker compose up -d +TAG=0.1.0 docker compose up -d ``` > Notice: Currently only the **Habana Driver 1.16.x** is supported for Gaudi. @@ -91,7 +91,7 @@ Find the corresponding [compose.yaml](./docker/xeon/compose.yaml). ```bash cd GenAIExamples/ChatQnA/docker/xeon/ -docker compose up -d +TAG=0.1.0 docker compose up -d ``` Refer to the [Xeon Guide](./docker/xeon/README.md) for more instructions on building docker images from source. @@ -100,7 +100,7 @@ Refer to the [Xeon Guide](./docker/xeon/README.md) for more instructions on buil ```bash cd GenAIExamples/ChatQnA/docker/gpu/ -docker compose up -d +TAG=0.1.0 docker compose up -d ``` Refer to the [NVIDIA GPU Guide](./docker/gpu/README.md) for more instructions on building docker images from source. diff --git a/ChatQnA/benchmark/single_gaudi/chatqna_mega_service_run.yaml b/ChatQnA/benchmark/single_gaudi/chatqna_mega_service_run.yaml index e255b17e7f..bc77cd04ba 100644 --- a/ChatQnA/benchmark/single_gaudi/chatqna_mega_service_run.yaml +++ b/ChatQnA/benchmark/single_gaudi/chatqna_mega_service_run.yaml @@ -30,7 +30,7 @@ spec: - envFrom: - configMapRef: name: qna-config - image: opea/chatqna:latest + image: opea/chatqna:latest.1.0 imagePullPolicy: IfNotPresent name: chatqna-backend-server-deploy args: null diff --git a/ChatQnA/benchmark/single_gaudi/dataprep-microservice_run.yaml b/ChatQnA/benchmark/single_gaudi/dataprep-microservice_run.yaml index de4d0716a7..24dcdf723e 100644 --- a/ChatQnA/benchmark/single_gaudi/dataprep-microservice_run.yaml +++ b/ChatQnA/benchmark/single_gaudi/dataprep-microservice_run.yaml @@ -38,7 +38,7 @@ spec: configMapKeyRef: name: qna-config key: INDEX_NAME - image: opea/dataprep-redis:latest + image: opea/dataprep-redis:latest.1.0 imagePullPolicy: IfNotPresent name: dataprep-deploy args: null diff --git a/ChatQnA/benchmark/single_gaudi/embedding-microservice_run.yaml b/ChatQnA/benchmark/single_gaudi/embedding-microservice_run.yaml index 2427872ffb..220762becb 100644 --- a/ChatQnA/benchmark/single_gaudi/embedding-microservice_run.yaml +++ b/ChatQnA/benchmark/single_gaudi/embedding-microservice_run.yaml @@ -30,7 +30,7 @@ spec: - envFrom: - configMapRef: name: qna-config - image: opea/embedding-tei:latest + image: opea/embedding-tei:latest.1.0 imagePullPolicy: IfNotPresent name: embedding-deploy args: null diff --git a/ChatQnA/benchmark/single_gaudi/llm-microservice_run.yaml b/ChatQnA/benchmark/single_gaudi/llm-microservice_run.yaml index 8afbc34673..107a5dee83 100644 --- a/ChatQnA/benchmark/single_gaudi/llm-microservice_run.yaml +++ b/ChatQnA/benchmark/single_gaudi/llm-microservice_run.yaml @@ -30,7 +30,7 @@ spec: - envFrom: - configMapRef: name: qna-config - image: opea/llm-tgi:latest + image: opea/llm-tgi:latest.1.0 imagePullPolicy: IfNotPresent name: llm-deploy args: null diff --git a/ChatQnA/benchmark/single_gaudi/reranking-microservice_run.yaml b/ChatQnA/benchmark/single_gaudi/reranking-microservice_run.yaml index 24fab63fe4..8205df01c3 100644 --- a/ChatQnA/benchmark/single_gaudi/reranking-microservice_run.yaml +++ b/ChatQnA/benchmark/single_gaudi/reranking-microservice_run.yaml @@ -30,7 +30,7 @@ spec: - envFrom: - configMapRef: name: qna-config - image: opea/reranking-tei:latest + image: opea/reranking-tei:latest.1.0 imagePullPolicy: IfNotPresent name: reranking-deploy args: null diff --git a/ChatQnA/benchmark/single_gaudi/retrieval-microservice_run.yaml b/ChatQnA/benchmark/single_gaudi/retrieval-microservice_run.yaml index e16505fcc6..a16648187b 100644 --- a/ChatQnA/benchmark/single_gaudi/retrieval-microservice_run.yaml +++ b/ChatQnA/benchmark/single_gaudi/retrieval-microservice_run.yaml @@ -38,7 +38,7 @@ spec: configMapKeyRef: name: qna-config key: INDEX_NAME - image: opea/retriever-redis:latest + image: opea/retriever-redis:latest.1.0 imagePullPolicy: IfNotPresent name: retriever-deploy args: null diff --git a/ChatQnA/benchmark/two_gaudi/chatqna_mega_service_run.yaml b/ChatQnA/benchmark/two_gaudi/chatqna_mega_service_run.yaml index e255b17e7f..bc77cd04ba 100644 --- a/ChatQnA/benchmark/two_gaudi/chatqna_mega_service_run.yaml +++ b/ChatQnA/benchmark/two_gaudi/chatqna_mega_service_run.yaml @@ -30,7 +30,7 @@ spec: - envFrom: - configMapRef: name: qna-config - image: opea/chatqna:latest + image: opea/chatqna:latest.1.0 imagePullPolicy: IfNotPresent name: chatqna-backend-server-deploy args: null diff --git a/ChatQnA/benchmark/two_gaudi/dataprep-microservice_run.yaml b/ChatQnA/benchmark/two_gaudi/dataprep-microservice_run.yaml index a9542d79de..a1d9c15ebd 100644 --- a/ChatQnA/benchmark/two_gaudi/dataprep-microservice_run.yaml +++ b/ChatQnA/benchmark/two_gaudi/dataprep-microservice_run.yaml @@ -38,7 +38,7 @@ spec: configMapKeyRef: name: qna-config key: INDEX_NAME - image: opea/dataprep-redis:latest + image: opea/dataprep-redis:latest.1.0 imagePullPolicy: IfNotPresent name: dataprep-deploy args: null diff --git a/ChatQnA/benchmark/two_gaudi/embedding-microservice_run.yaml b/ChatQnA/benchmark/two_gaudi/embedding-microservice_run.yaml index 2427872ffb..220762becb 100644 --- a/ChatQnA/benchmark/two_gaudi/embedding-microservice_run.yaml +++ b/ChatQnA/benchmark/two_gaudi/embedding-microservice_run.yaml @@ -30,7 +30,7 @@ spec: - envFrom: - configMapRef: name: qna-config - image: opea/embedding-tei:latest + image: opea/embedding-tei:latest.1.0 imagePullPolicy: IfNotPresent name: embedding-deploy args: null diff --git a/ChatQnA/benchmark/two_gaudi/llm-microservice_run.yaml b/ChatQnA/benchmark/two_gaudi/llm-microservice_run.yaml index 8afbc34673..107a5dee83 100644 --- a/ChatQnA/benchmark/two_gaudi/llm-microservice_run.yaml +++ b/ChatQnA/benchmark/two_gaudi/llm-microservice_run.yaml @@ -30,7 +30,7 @@ spec: - envFrom: - configMapRef: name: qna-config - image: opea/llm-tgi:latest + image: opea/llm-tgi:latest.1.0 imagePullPolicy: IfNotPresent name: llm-deploy args: null diff --git a/ChatQnA/benchmark/two_gaudi/reranking-microservice_run.yaml b/ChatQnA/benchmark/two_gaudi/reranking-microservice_run.yaml index 24fab63fe4..8205df01c3 100644 --- a/ChatQnA/benchmark/two_gaudi/reranking-microservice_run.yaml +++ b/ChatQnA/benchmark/two_gaudi/reranking-microservice_run.yaml @@ -30,7 +30,7 @@ spec: - envFrom: - configMapRef: name: qna-config - image: opea/reranking-tei:latest + image: opea/reranking-tei:latest.1.0 imagePullPolicy: IfNotPresent name: reranking-deploy args: null diff --git a/ChatQnA/benchmark/two_gaudi/retrieval-microservice_run.yaml b/ChatQnA/benchmark/two_gaudi/retrieval-microservice_run.yaml index e16505fcc6..a16648187b 100644 --- a/ChatQnA/benchmark/two_gaudi/retrieval-microservice_run.yaml +++ b/ChatQnA/benchmark/two_gaudi/retrieval-microservice_run.yaml @@ -38,7 +38,7 @@ spec: configMapKeyRef: name: qna-config key: INDEX_NAME - image: opea/retriever-redis:latest + image: opea/retriever-redis:latest.1.0 imagePullPolicy: IfNotPresent name: retriever-deploy args: null diff --git a/ChatQnA/chatqna.yaml b/ChatQnA/chatqna.yaml index 32cc433a7c..7a5b50e08c 100644 --- a/ChatQnA/chatqna.yaml +++ b/ChatQnA/chatqna.yaml @@ -12,14 +12,14 @@ opea_micro_services: dataprep-redis-service: host: ${DATAPREP_SERVICE_HOST_IP} ports: ${DATAPREP_SERVICE_PORT} - image: opea/dataprep-redis:latest + image: opea/dataprep-redis:latest.1.0 environment: REDIS_URL: ${REDIS_URL} INDEX_NAME: ${INDEX_NAME} tei-embedding-service: host: ${TEI_EMBEDDING_SERVICE_IP} ports: ${TEI_EMBEDDING_SERVICE_PORT} - image: opea/tei-gaudi:latest + image: opea/tei-gaudi:latest.1.0 volumes: - "./data:/data" runtime: habana @@ -33,17 +33,17 @@ opea_micro_services: embedding: host: ${EMBEDDING_SERVICE_HOST_IP} ports: ${EMBEDDING_SERVICE_PORT} - image: opea/embedding-tei:latest + image: opea/embedding-tei:latest.1.0 endpoint: /v1/embeddings retrieval: host: ${RETRIEVER_SERVICE_HOST_IP} ports: ${RETRIEVER_SERVICE_PORT} - image: opea/retriever-redis:latest + image: opea/retriever-redis:latest.1.0 endpoint: /v1/retrieval reranking: host: ${RERANK_SERVICE_HOST_IP} ports: ${RERANK_SERVICE_PORT} - image: opea/reranking-tei:latest + image: opea/reranking-tei:latest.1.0 endpoint: /v1/reranking tgi-service: host: ${TGI_SERVICE_IP} @@ -64,13 +64,13 @@ opea_micro_services: llm: host: ${LLM_SERVICE_HOST_IP} ports: ${LLM_SERVICE_PORT} - image: opea/llm-tgi:latest + image: opea/llm-tgi:latest.1.0 endpoint: /v1/chat/completions ui: host: ${UI_SERVICE_HOST_IP} ports: - "5173:5173" - image: opea/chatqna-ui:latest + image: opea/chatqna-ui:latest.1.0 environment: - CHAT_BASE_URL=${BACKEND_SERVICE_ENDPOINT} - UPLOAD_FILE_BASE_URL=${DATAPREP_SERVICE_ENDPOINT} @@ -78,7 +78,7 @@ opea_micro_services: opea_mega_service: host: ${MEGA_SERVICE_HOST_IP} ports: ${MEGA_SERVICE_PORT} - image: opea/chatqna:latest + image: opea/chatqna:latest.1.0 endpoint: /v1/chatqna mega_flow: - embedding >> retrieval >> reranking >> llm diff --git a/ChatQnA/docker/aipc/README.md b/ChatQnA/docker/aipc/README.md index 5396421c07..73c0e304a2 100644 --- a/ChatQnA/docker/aipc/README.md +++ b/ChatQnA/docker/aipc/README.md @@ -14,19 +14,19 @@ cd GenAIComps ### 1. Build Embedding Image ```bash -docker build --no-cache -t opea/embedding-tei:latest -f comps/embeddings/langchain/docker/Dockerfile . +docker build --no-cache -t opea/embedding-tei:latest.1.0 -f comps/embeddings/langchain/docker/Dockerfile . ``` ### 2. Build Retriever Image ```bash -docker build --no-cache -t opea/retriever-redis:latest -f comps/retrievers/langchain/redis/docker/Dockerfile . +docker build --no-cache -t opea/retriever-redis:latest.1.0 -f comps/retrievers/langchain/redis/docker/Dockerfile . ``` ### 3. Build Rerank Image ```bash -docker build --no-cache -t opea/reranking-tei:latest -f comps/reranks/tei/docker/Dockerfile . +docker build --no-cache -t opea/reranking-tei:latest.1.0 -f comps/reranks/tei/docker/Dockerfile . ``` ### 4. Build LLM Image @@ -34,13 +34,13 @@ docker build --no-cache -t opea/reranking-tei:latest -f comps/reranks/tei/docker We use [Ollama](https://ollama.com/) as our LLM service for AIPC. Please pre-download Ollama on your PC. ```bash -docker build --no-cache -t opea/llm-ollama:latest -f comps/llms/text-generation/ollama/Dockerfile . +docker build --no-cache -t opea/llm-ollama:latest.1.0 -f comps/llms/text-generation/ollama/Dockerfile . ``` ### 5. Build Dataprep Image ```bash -docker build --no-cache -t opea/dataprep-redis:latest -f comps/dataprep/redis/langchain/docker/Dockerfile . +docker build --no-cache -t opea/dataprep-redis:latest.1.0 -f comps/dataprep/redis/langchain/docker/Dockerfile . cd .. ``` @@ -51,7 +51,7 @@ To construct the Mega Service, we utilize the [GenAIComps](https://github.com/op ```bash git clone https://github.com/opea-project/GenAIExamples.git cd GenAIExamples/ChatQnA/docker -docker build --no-cache -t opea/chatqna:latest -f Dockerfile . +docker build --no-cache -t opea/chatqna:latest.1.0 -f Dockerfile . cd ../../.. ``` @@ -61,19 +61,19 @@ Build frontend Docker image via below command: ```bash cd GenAIExamples/ChatQnA/docker/ui/ -docker build --no-cache -t opea/chatqna-ui:latest -f ./docker/Dockerfile . +docker build --no-cache -t opea/chatqna-ui:latest.1.0 -f ./docker/Dockerfile . cd ../../../.. ``` Then run the command `docker images`, you will have the following 7 Docker Images: -1. `opea/dataprep-redis:latest` -2. `opea/embedding-tei:latest` -3. `opea/retriever-redis:latest` -4. `opea/reranking-tei:latest` -5. `opea/llm-ollama:latest` -6. `opea/chatqna:latest` -7. `opea/chatqna-ui:latest` +1. `opea/dataprep-redis:latest.1.0` +2. `opea/embedding-tei:latest.1.0` +3. `opea/retriever-redis:latest.1.0` +4. `opea/reranking-tei:latest.1.0` +5. `opea/llm-ollama:latest.1.0` +6. `opea/chatqna:latest.1.0` +7. `opea/chatqna-ui:latest.1.0` ## πŸš€ Start Microservices @@ -160,7 +160,7 @@ Note: Please replace with `host_ip` with you external IP address, do not use loc ```bash cd GenAIExamples/ChatQnA/docker/aipc/ -docker compose up -d +TAG=0.1.0 docker compose up -d # let ollama service runs # e.g. ollama run llama3 diff --git a/ChatQnA/docker/gaudi/README.md b/ChatQnA/docker/gaudi/README.md index 75bd412d57..ef03b8dfe2 100644 --- a/ChatQnA/docker/gaudi/README.md +++ b/ChatQnA/docker/gaudi/README.md @@ -16,19 +16,19 @@ cd GenAIComps ### 2. Build Embedding Image ```bash -docker build --no-cache -t opea/embedding-tei:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/embeddings/langchain/docker/Dockerfile . +docker build --no-cache -t opea/embedding-tei:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/embeddings/langchain/docker/Dockerfile . ``` ### 3. Build Retriever Image ```bash -docker build --no-cache -t opea/retriever-redis:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/retrievers/langchain/redis/docker/Dockerfile . +docker build --no-cache -t opea/retriever-redis:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/retrievers/langchain/redis/docker/Dockerfile . ``` ### 4. Build Rerank Image ```bash -docker build --no-cache -t opea/reranking-tei:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/reranks/tei/docker/Dockerfile . +docker build --no-cache -t opea/reranking-tei:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/reranks/tei/docker/Dockerfile . ``` ### 5. Build LLM Image @@ -38,7 +38,7 @@ You can use different LLM serving solutions, choose one of following four option #### 5.1 Use TGI ```bash -docker build --no-cache -t opea/llm-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . +docker build --no-cache -t opea/llm-tgi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . ``` #### 5.2 Use VLLM @@ -46,13 +46,13 @@ docker build --no-cache -t opea/llm-tgi:latest --build-arg https_proxy=$https_pr Build vllm docker. ```bash -docker build --no-cache -t opea/llm-vllm-hpu:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/vllm/docker/Dockerfile.hpu . +docker build --no-cache -t opea/llm-vllm-hpu:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/vllm/docker/Dockerfile.hpu . ``` Build microservice docker. ```bash -docker build --no-cache -t opea/llm-vllm:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/vllm/docker/Dockerfile.microservice . +docker build --no-cache -t opea/llm-vllm:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/vllm/docker/Dockerfile.microservice . ``` #### 5.3 Use VLLM-on-Ray @@ -60,19 +60,19 @@ docker build --no-cache -t opea/llm-vllm:latest --build-arg https_proxy=$https_p Build vllm-on-ray docker. ```bash -docker build --no-cache -t opea/llm-vllm-ray-hpu:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/vllm-ray/docker/Dockerfile.vllmray . +docker build --no-cache -t opea/llm-vllm-ray-hpu:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/vllm-ray/docker/Dockerfile.vllmray . ``` Build microservice docker. ```bash -docker build --no-cache -t opea/llm-vllm-ray:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/vllm-ray/docker/Dockerfile.microservice . +docker build --no-cache -t opea/llm-vllm-ray:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/vllm-ray/docker/Dockerfile.microservice . ``` ### 6. Build Dataprep Image ```bash -docker build --no-cache -t opea/dataprep-redis:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/dataprep/redis/langchain/docker/Dockerfile . +docker build --no-cache -t opea/dataprep-redis:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/dataprep/redis/langchain/docker/Dockerfile . ``` ### 7. Build TEI Gaudi Image @@ -82,7 +82,7 @@ Since a TEI Gaudi Docker image hasn't been published, we'll need to build it fro ```bash git clone https://github.com/huggingface/tei-gaudi cd tei-gaudi/ -docker build --no-cache -f Dockerfile-hpu -t opea/tei-gaudi:latest . +docker build --no-cache -f Dockerfile-hpu -t opea/tei-gaudi:latest.1.0 . cd ../.. ``` @@ -93,7 +93,7 @@ To construct the Mega Service, we utilize the [GenAIComps](https://github.com/op ```bash git clone https://github.com/opea-project/GenAIExamples.git cd GenAIExamples/ChatQnA/docker -docker build --no-cache -t opea/chatqna:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . +docker build --no-cache -t opea/chatqna:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . cd ../../.. ``` @@ -102,7 +102,7 @@ If you want to enable guardrails microservice in the pipeline, please use the be ```bash git clone https://github.com/opea-project/GenAIExamples.git cd GenAIExamples/ChatQnA/docker -docker build --no-cache -t opea/chatqna-guardrails:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile_guardrails . +docker build --no-cache -t opea/chatqna-guardrails:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile_guardrails . cd ../../.. ``` @@ -112,7 +112,7 @@ Construct the frontend Docker image using the command below: ```bash cd GenAIExamples/ChatQnA/docker/ui/ -docker build --no-cache -t opea/chatqna-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . +docker build --no-cache -t opea/chatqna-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . cd ../../../.. ``` @@ -124,7 +124,7 @@ Build frontend Docker image that enables Conversational experience with ChatQnA ```bash cd GenAIExamples/ChatQnA/docker/ui/ -docker build --no-cache -t opea/chatqna-conversation-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile.react . +docker build --no-cache -t opea/chatqna-conversation-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile.react . cd ../../../.. ``` @@ -134,28 +134,28 @@ To fortify AI initiatives in production, Guardrails microservice can secure mode ```bash cd GenAIExamples/ChatQnA/docker -docker build -t opea/guardrails-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/guardrails/llama_guard/docker/Dockerfile . +docker build -t opea/guardrails-tgi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/guardrails/llama_guard/docker/Dockerfile . cd ../../.. ``` Then run the command `docker images`, you will have the following 8 Docker Images: -1. `opea/embedding-tei:latest` -2. `opea/retriever-redis:latest` -3. `opea/reranking-tei:latest` -4. `opea/llm-tgi:latest` or `opea/llm-vllm:latest` or `opea/llm-vllm-ray:latest` -5. `opea/tei-gaudi:latest` -6. `opea/dataprep-redis:latest` -7. `opea/chatqna:latest` or `opea/chatqna-guardrails:latest` -8. `opea/chatqna-ui:latest` +1. `opea/embedding-tei:latest.1.0` +2. `opea/retriever-redis:latest.1.0` +3. `opea/reranking-tei:latest.1.0` +4. `opea/llm-tgi:latest.1.0` or `opea/llm-vllm:latest.1.0` or `opea/llm-vllm-ray:latest.1.0` +5. `opea/tei-gaudi:latest.1.0` +6. `opea/dataprep-redis:latest.1.0` +7. `opea/chatqna:latest.1.0` or `opea/chatqna-guardrails:latest.1.0` +8. `opea/chatqna-ui:latest.1.0` If Conversation React UI is built, you will find one more image: -9. `opea/chatqna-conversation-ui:latest` +9. `opea/chatqna-conversation-ui:latest.1.0` If Guardrails docker image is built, you will find one more image: -10. `opea/guardrails-tgi:latest` +10. `opea/guardrails-tgi:latest.1.0` ## πŸš€ Start MicroServices and MegaService @@ -211,26 +211,26 @@ cd GenAIExamples/ChatQnA/docker/gaudi/ If use tgi for llm backend. ```bash -docker compose -f compose.yaml up -d +TAG=0.1.0 docker compose -f compose.yaml up -d ``` If use vllm for llm backend. ```bash -docker compose -f compose_vllm.yaml up -d +TAG=0.1.0 docker compose -f compose_vllm.yaml up -d ``` If use vllm-on-ray for llm backend. ```bash -docker compose -f compose_vllm_ray.yaml up -d +TAG=0.1.0 docker compose -f compose_vllm_ray.yaml up -d ``` If you want to enable guardrails microservice in the pipeline, please follow the below command instead: ```bash cd GenAIExamples/ChatQnA/docker/gaudi/ -docker compose -f compose_guardrails.yaml up -d +TAG=0.1.0 docker compose -f compose_guardrails.yaml up -d ``` > **_NOTE:_** Users need at least two Gaudi cards to run the ChatQnA successfully. @@ -403,7 +403,7 @@ To access the frontend, open the following URL in your browser: http://{host_ip} ```yaml chaqna-gaudi-ui-server: - image: opea/chatqna-ui:latest + image: opea/chatqna-ui:latest.1.0 ... ports: - "80:5173" @@ -421,7 +421,7 @@ To access the Conversational UI (react based) frontend, modify the UI service in ```yaml chaqna-gaudi-conversation-ui-server: - image: opea/chatqna-conversation-ui:latest + image: opea/chatqna-conversation-ui:latest.1.0 container_name: chatqna-gaudi-conversation-ui-server environment: - APP_BACKEND_SERVICE_ENDPOINT=${BACKEND_SERVICE_ENDPOINT} @@ -438,7 +438,7 @@ Once the services are up, open the following URL in your browser: http://{host_i ```yaml chaqna-gaudi-conversation-ui-server: - image: opea/chatqna-conversation-ui:latest + image: opea/chatqna-conversation-ui:latest.1.0 ... ports: - "80:80" diff --git a/ChatQnA/docker/gaudi/how_to_validate_service.md b/ChatQnA/docker/gaudi/how_to_validate_service.md index 0e58491eb8..ca85a6afce 100644 --- a/ChatQnA/docker/gaudi/how_to_validate_service.md +++ b/ChatQnA/docker/gaudi/how_to_validate_service.md @@ -17,7 +17,7 @@ start the docker containers ``` cd ./GenAIExamples/ChatQnA/docker/gaudi -docker compose up -d +TAG=0.1.0 docker compose up -d ``` Check the start up log by `docker compose -f ./docker/gaudi/compose.yaml logs`. @@ -49,15 +49,15 @@ Here is the output: ``` CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES -28d9a5570246 opea/chatqna-ui:latest "docker-entrypoint.s…" 2 minutes ago Up 2 minutes 0.0.0.0:5173->5173/tcp, :::5173->5173/tcp chatqna-gaudi-ui-server -bee1132464cd opea/chatqna:latest "python chatqna.py" 2 minutes ago Up 2 minutes 0.0.0.0:8888->8888/tcp, :::8888->8888/tcp chatqna-gaudi-backend-server -f810f3b4d329 opea/embedding-tei:latest "python embedding_te…" 2 minutes ago Up 2 minutes 0.0.0.0:6000->6000/tcp, :::6000->6000/tcp embedding-tei-server -325236a01f9b opea/llm-tgi:latest "python llm.py" 2 minutes ago Up 2 minutes 0.0.0.0:9000->9000/tcp, :::9000->9000/tcp llm-tgi-gaudi-server -2fa17d84605f opea/dataprep-redis:latest "python prepare_doc_…" 2 minutes ago Up 2 minutes 0.0.0.0:6007->6007/tcp, :::6007->6007/tcp dataprep-redis-server -69e1fb59e92c opea/retriever-redis:latest "/home/user/comps/re…" 2 minutes ago Up 2 minutes 0.0.0.0:7000->7000/tcp, :::7000->7000/tcp retriever-redis-server -313b9d14928a opea/reranking-tei:latest "python reranking_te…" 2 minutes ago Up 2 minutes 0.0.0.0:8000->8000/tcp, :::8000->8000/tcp reranking-tei-gaudi-server +28d9a5570246 opea/chatqna-ui:latest.1.0 "docker-entrypoint.s…" 2 minutes ago Up 2 minutes 0.0.0.0:5173->5173/tcp, :::5173->5173/tcp chatqna-gaudi-ui-server +bee1132464cd opea/chatqna:latest.1.0 "python chatqna.py" 2 minutes ago Up 2 minutes 0.0.0.0:8888->8888/tcp, :::8888->8888/tcp chatqna-gaudi-backend-server +f810f3b4d329 opea/embedding-tei:latest.1.0 "python embedding_te…" 2 minutes ago Up 2 minutes 0.0.0.0:6000->6000/tcp, :::6000->6000/tcp embedding-tei-server +325236a01f9b opea/llm-tgi:latest.1.0 "python llm.py" 2 minutes ago Up 2 minutes 0.0.0.0:9000->9000/tcp, :::9000->9000/tcp llm-tgi-gaudi-server +2fa17d84605f opea/dataprep-redis:latest.1.0 "python prepare_doc_…" 2 minutes ago Up 2 minutes 0.0.0.0:6007->6007/tcp, :::6007->6007/tcp dataprep-redis-server +69e1fb59e92c opea/retriever-redis:latest.1.0 "/home/user/comps/re…" 2 minutes ago Up 2 minutes 0.0.0.0:7000->7000/tcp, :::7000->7000/tcp retriever-redis-server +313b9d14928a opea/reranking-tei:latest.1.0 "python reranking_te…" 2 minutes ago Up 2 minutes 0.0.0.0:8000->8000/tcp, :::8000->8000/tcp reranking-tei-gaudi-server 05c40b636239 ghcr.io/huggingface/tgi-gaudi:1.2.1 "text-generation-lau…" 2 minutes ago Exited (1) About a minute ago tgi-gaudi-server -174bd43fa6b5 opea/tei-gaudi:latest "text-embeddings-rou…" 2 minutes ago Up 2 minutes 0.0.0.0:8090->80/tcp, :::8090->80/tcp tei-embedding-gaudi-server +174bd43fa6b5 opea/tei-gaudi:latest.1.0 "text-embeddings-rou…" 2 minutes ago Up 2 minutes 0.0.0.0:8090->80/tcp, :::8090->80/tcp tei-embedding-gaudi-server 74084469aa33 redis/redis-stack:7.2.0-v9 "/entrypoint.sh" 2 minutes ago Up 2 minutes 0.0.0.0:6379->6379/tcp, :::6379->6379/tcp, 0.0.0.0:8001->8001/tcp, :::8001->8001/tcp redis-vector-db 88399dbc9e43 ghcr.io/huggingface/text-embeddings-inference:cpu-1.5 "text-embeddings-rou…" 2 minutes ago Up 2 minutes 0.0.0.0:8808->80/tcp, :::8808->80/tcp tei-reranking-gaudi-server ``` @@ -149,7 +149,7 @@ Set the LLM_MODEL_ID then restart the containers. Also you can check overall logs with the following command, where the compose.yaml is the mega service docker-compose configuration file. ``` -docker compose -f ./docker-composer/gaudi/compose.yaml logs +TAG=0.1.0 docker compose -f ./docker-composer/gaudi/compose.yaml logs ``` ## 4. Check each micro service used by the Mega Service diff --git a/ChatQnA/docker/gpu/README.md b/ChatQnA/docker/gpu/README.md index 41c2aef329..d70decb864 100644 --- a/ChatQnA/docker/gpu/README.md +++ b/ChatQnA/docker/gpu/README.md @@ -16,31 +16,31 @@ cd GenAIComps ### 2. Build Embedding Image ```bash -docker build --no-cache -t opea/embedding-tei:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/embeddings/langchain/docker/Dockerfile . +docker build --no-cache -t opea/embedding-tei:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/embeddings/langchain/docker/Dockerfile . ``` ### 3. Build Retriever Image ```bash -docker build --no-cache -t opea/retriever-redis:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/retrievers/langchain/redis/docker/Dockerfile . +docker build --no-cache -t opea/retriever-redis:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/retrievers/langchain/redis/docker/Dockerfile . ``` ### 4. Build Rerank Image ```bash -docker build --no-cache -t opea/reranking-tei:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/reranks/tei/docker/Dockerfile . +docker build --no-cache -t opea/reranking-tei:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/reranks/tei/docker/Dockerfile . ``` ### 5. Build LLM Image ```bash -docker build --no-cache -t opea/llm-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . +docker build --no-cache -t opea/llm-tgi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . ``` ### 6. Build Dataprep Image ```bash -docker build --no-cache -t opea/dataprep-redis:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/dataprep/redis/langchain/docker/Dockerfile . +docker build --no-cache -t opea/dataprep-redis:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/dataprep/redis/langchain/docker/Dockerfile . ``` ### 7. Build MegaService Docker Image @@ -50,7 +50,7 @@ To construct the Mega Service, we utilize the [GenAIComps](https://github.com/op ```bash git clone https://github.com/opea-project/GenAIExamples.git cd GenAIExamples/ChatQnA/docker -docker build --no-cache -t opea/chatqna:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . +docker build --no-cache -t opea/chatqna:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . cd ../../.. ``` @@ -60,7 +60,7 @@ Construct the frontend Docker image using the command below: ```bash cd GenAIExamples/ChatQnA/docker/ui/ -docker build --no-cache -t opea/chatqna-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . +docker build --no-cache -t opea/chatqna-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . cd ../../../.. ``` @@ -70,20 +70,20 @@ Construct the frontend Docker image using the command below: ```bash cd GenAIExamples/ChatQnA/docker/ui/ -docker build --no-cache -t opea/chatqna-react-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile.react . +docker build --no-cache -t opea/chatqna-react-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile.react . cd ../../../.. ``` Then run the command `docker images`, you will have the following 7 Docker Images: -1. `opea/embedding-tei:latest` -2. `opea/retriever-redis:latest` -3. `opea/reranking-tei:latest` -4. `opea/llm-tgi:latest` -5. `opea/dataprep-redis:latest` -6. `opea/chatqna:latest` -7. `opea/chatqna-ui:latest` -8. `opea/chatqna-react-ui:latest` +1. `opea/embedding-tei:latest.1.0` +2. `opea/retriever-redis:latest.1.0` +3. `opea/reranking-tei:latest.1.0` +4. `opea/llm-tgi:latest.1.0` +5. `opea/dataprep-redis:latest.1.0` +6. `opea/chatqna:latest.1.0` +7. `opea/chatqna-ui:latest.1.0` +8. `opea/chatqna-react-ui:latest.1.0` ## πŸš€ Start MicroServices and MegaService @@ -121,7 +121,7 @@ Note: Please replace with `host_ip` with you external IP address, do **NOT** use ```bash cd GenAIExamples/ChatQnA/docker/gpu/ -docker compose up -d +TAG=0.1.0 docker compose up -d ``` ### Validate MicroServices and MegaService @@ -260,7 +260,7 @@ To access the frontend, open the following URL in your browser: http://{host_ip} ```yaml chaqna-ui-server: - image: opea/chatqna-ui:latest + image: opea/chatqna-ui:latest.1.0 ... ports: - "80:5173" @@ -272,7 +272,7 @@ To access the Conversational UI (react based) frontend, modify the UI service in ```yaml chaqna-xeon-conversation-ui-server: - image: opea/chatqna-conversation-ui:latest + image: opea/chatqna-conversation-ui:latest.1.0 container_name: chatqna-xeon-conversation-ui-server environment: - APP_BACKEND_SERVICE_ENDPOINT=${BACKEND_SERVICE_ENDPOINT} @@ -289,7 +289,7 @@ Once the services are up, open the following URL in your browser: http://{host_i ```yaml chaqna-xeon-conversation-ui-server: - image: opea/chatqna-conversation-ui:latest + image: opea/chatqna-conversation-ui:latest.1.0 ... ports: - "80:80" diff --git a/ChatQnA/docker/xeon/README.md b/ChatQnA/docker/xeon/README.md index b607e8532e..aba309f606 100644 --- a/ChatQnA/docker/xeon/README.md +++ b/ChatQnA/docker/xeon/README.md @@ -69,19 +69,19 @@ cd GenAIComps ### 1. Build Embedding Image ```bash -docker build --no-cache -t opea/embedding-tei:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/embeddings/langchain/docker/Dockerfile . +docker build --no-cache -t opea/embedding-tei:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/embeddings/langchain/docker/Dockerfile . ``` ### 2. Build Retriever Image ```bash -docker build --no-cache -t opea/retriever-redis:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/retrievers/langchain/redis/docker/Dockerfile . +docker build --no-cache -t opea/retriever-redis:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/retrievers/langchain/redis/docker/Dockerfile . ``` ### 3. Build Rerank Image ```bash -docker build --no-cache -t opea/reranking-tei:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/reranks/tei/docker/Dockerfile . +docker build --no-cache -t opea/reranking-tei:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/reranks/tei/docker/Dockerfile . ``` ### 4. Build LLM Image @@ -89,7 +89,7 @@ docker build --no-cache -t opea/reranking-tei:latest --build-arg https_proxy=$ht #### Use TGI as backend ```bash -docker build --no-cache -t opea/llm-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . +docker build --no-cache -t opea/llm-tgi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . ``` #### Use vLLM as backend @@ -99,20 +99,20 @@ Build vLLM docker. ```bash git clone https://github.com/vllm-project/vllm.git cd ./vllm/ -docker build --no-cache -t opea/vllm:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile.cpu . +docker build --no-cache -t opea/vllm:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile.cpu . cd .. ``` Build microservice. ```bash -docker build --no-cache -t opea/llm-vllm:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/vllm/docker/Dockerfile.microservice . +docker build --no-cache -t opea/llm-vllm:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/vllm/docker/Dockerfile.microservice . ``` ### 5. Build Dataprep Image ```bash -docker build --no-cache -t opea/dataprep-redis:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/dataprep/redis/langchain/docker/Dockerfile . +docker build --no-cache -t opea/dataprep-redis:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/dataprep/redis/langchain/docker/Dockerfile . cd .. ``` @@ -123,7 +123,7 @@ To construct the Mega Service, we utilize the [GenAIComps](https://github.com/op ```bash git clone https://github.com/opea-project/GenAIExamples.git cd GenAIExamples/ChatQnA/docker -docker build --no-cache -t opea/chatqna:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . +docker build --no-cache -t opea/chatqna:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . cd ../../.. ``` @@ -133,7 +133,7 @@ Build frontend Docker image via below command: ```bash cd GenAIExamples/ChatQnA/docker/ui/ -docker build --no-cache -t opea/chatqna-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . +docker build --no-cache -t opea/chatqna-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . cd ../../../.. ``` @@ -145,19 +145,19 @@ Build frontend Docker image that enables Conversational experience with ChatQnA ```bash cd GenAIExamples/ChatQnA/docker/ui/ -docker build --no-cache -t opea/chatqna-conversation-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile.react . +docker build --no-cache -t opea/chatqna-conversation-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile.react . cd ../../../.. ``` Then run the command `docker images`, you will have the following 7 Docker Images: -1. `opea/dataprep-redis:latest` -2. `opea/embedding-tei:latest` -3. `opea/retriever-redis:latest` -4. `opea/reranking-tei:latest` -5. `opea/llm-tgi:latest` or `opea/llm-vllm:latest` -6. `opea/chatqna:latest` -7. `opea/chatqna-ui:latest` +1. `opea/dataprep-redis:latest.1.0` +2. `opea/embedding-tei:latest.1.0` +3. `opea/retriever-redis:latest.1.0` +4. `opea/reranking-tei:latest.1.0` +5. `opea/llm-tgi:latest.1.0` or `opea/llm-vllm:latest.1.0` +6. `opea/chatqna:latest.1.0` +7. `opea/chatqna-ui:latest.1.0` ## πŸš€ Start Microservices @@ -226,13 +226,13 @@ cd GenAIExamples/ChatQnA/docker/xeon/ If use TGI backend. ```bash -docker compose -f compose.yaml up -d +TAG=0.1.0 docker compose -f compose.yaml up -d ``` If use vLLM backend. ```bash -docker compose -f compose_vllm.yaml up -d +TAG=0.1.0 docker compose -f compose_vllm.yaml up -d ``` ### Validate Microservices @@ -385,7 +385,7 @@ To access the frontend, open the following URL in your browser: http://{host_ip} ```yaml chaqna-gaudi-ui-server: - image: opea/chatqna-ui:latest + image: opea/chatqna-ui:latest.1.0 ... ports: - "80:5173" @@ -397,7 +397,7 @@ To access the Conversational UI (react based) frontend, modify the UI service in ```yaml chaqna-xeon-conversation-ui-server: - image: opea/chatqna-conversation-ui:latest + image: opea/chatqna-conversation-ui:latest.1.0 container_name: chatqna-xeon-conversation-ui-server environment: - APP_BACKEND_SERVICE_ENDPOINT=${BACKEND_SERVICE_ENDPOINT} @@ -414,7 +414,7 @@ Once the services are up, open the following URL in your browser: http://{host_i ```yaml chaqna-gaudi-conversation-ui-server: - image: opea/chatqna-conversation-ui:latest + image: opea/chatqna-conversation-ui:latest.1.0 ... ports: - "80:80" diff --git a/ChatQnA/docker/xeon/README_qdrant.md b/ChatQnA/docker/xeon/README_qdrant.md index fd28d306ab..87a96debd9 100644 --- a/ChatQnA/docker/xeon/README_qdrant.md +++ b/ChatQnA/docker/xeon/README_qdrant.md @@ -73,31 +73,31 @@ cd GenAIComps ### 1. Build Embedding Image ```bash -docker build --no-cache -t opea/embedding-tei:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/embeddings/langchain/docker/Dockerfile . +docker build --no-cache -t opea/embedding-tei:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/embeddings/langchain/docker/Dockerfile . ``` ### 2. Build Retriever Image ```bash -docker build --no-cache -t opea/retriever-qdrant:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/retrievers/haystack/qdrant/docker/Dockerfile . +docker build --no-cache -t opea/retriever-qdrant:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/retrievers/haystack/qdrant/docker/Dockerfile . ``` ### 3. Build Rerank Image ```bash -docker build --no-cache -t opea/reranking-tei:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/reranks/tei/docker/Dockerfile . +docker build --no-cache -t opea/reranking-tei:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/reranks/tei/docker/Dockerfile . ``` ### 4. Build LLM Image ```bash -docker build --no-cache -t opea/llm-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . +docker build --no-cache -t opea/llm-tgi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . ``` ### 5. Build Dataprep Image ```bash -docker build --no-cache -t opea/dataprep-qdrant:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/dataprep/qdrant/docker/Dockerfile . +docker build --no-cache -t opea/dataprep-qdrant:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/dataprep/qdrant/docker/Dockerfile . cd .. ``` @@ -108,7 +108,7 @@ To construct the Mega Service, we utilize the [GenAIComps](https://github.com/op ```bash git clone https://github.com/opea-project/GenAIExamples.git cd GenAIExamples/ChatQnA/docker -docker build --no-cache -t opea/chatqna:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . +docker build --no-cache -t opea/chatqna:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . cd ../../.. ``` @@ -118,7 +118,7 @@ Build frontend Docker image via below command: ```bash cd GenAIExamples/ChatQnA/docker/ui/ -docker build --no-cache -t opea/chatqna-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . +docker build --no-cache -t opea/chatqna-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . cd ../../../.. ``` @@ -132,19 +132,19 @@ Build frontend Docker image that enables Conversational experience with ChatQnA cd GenAIExamples/ChatQnA/docker/ui/ export BACKEND_SERVICE_ENDPOINT="http://${host_ip}:8912/v1/chatqna" export DATAPREP_SERVICE_ENDPOINT="http://${host_ip}:6043/v1/dataprep" -docker build --no-cache -t opea/chatqna-conversation-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy --build-arg BACKEND_SERVICE_ENDPOINT=$BACKEND_SERVICE_ENDPOINT --build-arg DATAPREP_SERVICE_ENDPOINT=$DATAPREP_SERVICE_ENDPOINT -f ./docker/Dockerfile.react . +docker build --no-cache -t opea/chatqna-conversation-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy --build-arg BACKEND_SERVICE_ENDPOINT=$BACKEND_SERVICE_ENDPOINT --build-arg DATAPREP_SERVICE_ENDPOINT=$DATAPREP_SERVICE_ENDPOINT -f ./docker/Dockerfile.react . cd ../../../.. ``` Then run the command `docker images`, you will have the following 7 Docker Images: -1. `opea/dataprep-qdrant:latest` -2. `opea/embedding-tei:latest` -3. `opea/retriever-qdrant:latest` -4. `opea/reranking-tei:latest` -5. `opea/llm-tgi:latest` -6. `opea/chatqna:latest` -7. `opea/chatqna-ui:latest` +1. `opea/dataprep-qdrant:latest.1.0` +2. `opea/embedding-tei:latest.1.0` +3. `opea/retriever-qdrant:latest.1.0` +4. `opea/reranking-tei:latest.1.0` +5. `opea/llm-tgi:latest.1.0` +6. `opea/chatqna:latest.1.0` +7. `opea/chatqna-ui:latest.1.0` ## πŸš€ Start Microservices @@ -205,7 +205,7 @@ Note: Please replace with `host_ip` with you external IP address, do not use loc ```bash cd GenAIExamples/ChatQnA/docker/xeon/ -docker compose -f compose_qdrant.yaml up -d +TAG=0.1.0 docker compose -f compose_qdrant.yaml up -d ``` ### Validate Microservices @@ -316,7 +316,7 @@ To access the frontend, open the following URL in your browser: http://{host_ip} ```yaml chaqna-gaudi-ui-server: - image: opea/chatqna-ui:latest + image: opea/chatqna-ui:latest.1.0 ... ports: - "80:5173" @@ -328,7 +328,7 @@ To access the Conversational UI frontend, open the following URL in your browser ```yaml chaqna-xeon-conversation-ui-server: - image: opea/chatqna-conversation-ui:latest + image: opea/chatqna-conversation-ui:latest.1.0 ... ports: - "80:80" diff --git a/ChatQnA/kubernetes/README.md b/ChatQnA/kubernetes/README.md index 952203e5cd..c557dd65e6 100644 --- a/ChatQnA/kubernetes/README.md +++ b/ChatQnA/kubernetes/README.md @@ -16,18 +16,18 @@ The ChatQnA uses the below prebuilt images if you choose a Xeon deployment - redis-vector-db: redis/redis-stack:7.2.0-v9 - tei_embedding_service: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5 -- embedding: opea/embedding-tei:latest -- retriever: opea/retriever-redis:latest +- embedding: opea/embedding-tei:latest.1.0 +- retriever: opea/retriever-redis:latest.1.0 - tei_xeon_service: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5 -- reranking: opea/reranking-tei:latest +- reranking: opea/reranking-tei:latest.1.0 - tgi-service: ghcr.io/huggingface/text-generation-inference:2.2.0 -- llm: opea/llm-tgi:latest -- chaqna-xeon-backend-server: opea/chatqna:latest +- llm: opea/llm-tgi:latest.1.0 +- chaqna-xeon-backend-server: opea/chatqna:latest.1.0 Should you desire to use the Gaudi accelerator, two alternate images are used for the embedding and llm services. For Gaudi: -- tei-embedding-service: opea/tei-gaudi:latest +- tei-embedding-service: opea/tei-gaudi:latest.1.0 - tgi-service: ghcr.io/huggingface/tgi-gaudi:1.2.1 > [NOTE] diff --git a/ChatQnA/kubernetes/manifests/gaudi/chatqna.yaml b/ChatQnA/kubernetes/manifests/gaudi/chatqna.yaml index fbd6fabac8..fed16acd39 100644 --- a/ChatQnA/kubernetes/manifests/gaudi/chatqna.yaml +++ b/ChatQnA/kubernetes/manifests/gaudi/chatqna.yaml @@ -499,7 +499,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/dataprep-redis:latest" + image: "opea/dataprep-redis:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: data-prep @@ -557,7 +557,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/embedding-tei:latest" + image: "opea/embedding-tei:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: embedding-usvc @@ -615,7 +615,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/llm-tgi:latest" + image: "opea/llm-tgi:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: llm-uservice @@ -753,7 +753,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/reranking-tei:latest" + image: "opea/reranking-tei:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: reranking-usvc @@ -811,7 +811,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/retriever-redis:latest" + image: "opea/retriever-redis:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: retriever-usvc @@ -1065,7 +1065,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/chatqna:latest" + image: "opea/chatqna:latest.1.0" imagePullPolicy: IfNotPresent volumeMounts: - mountPath: /tmp diff --git a/ChatQnA/kubernetes/manifests/xeon/chatqna.yaml b/ChatQnA/kubernetes/manifests/xeon/chatqna.yaml index a99584d010..2a925570a1 100644 --- a/ChatQnA/kubernetes/manifests/xeon/chatqna.yaml +++ b/ChatQnA/kubernetes/manifests/xeon/chatqna.yaml @@ -500,7 +500,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/dataprep-redis:latest" + image: "opea/dataprep-redis:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: data-prep @@ -558,7 +558,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/embedding-tei:latest" + image: "opea/embedding-tei:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: embedding-usvc @@ -616,7 +616,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/llm-tgi:latest" + image: "opea/llm-tgi:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: llm-uservice @@ -754,7 +754,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/reranking-tei:latest" + image: "opea/reranking-tei:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: reranking-usvc @@ -812,7 +812,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/retriever-redis:latest" + image: "opea/retriever-redis:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: retriever-usvc @@ -1064,7 +1064,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/chatqna:latest" + image: "opea/chatqna:latest.1.0" imagePullPolicy: IfNotPresent volumeMounts: - mountPath: /tmp diff --git a/CodeGen/README.md b/CodeGen/README.md index d9b399a35c..ca3ad7058d 100644 --- a/CodeGen/README.md +++ b/CodeGen/README.md @@ -27,7 +27,7 @@ Currently we support two ways of deploying ChatQnA services with docker compose: 1. Start services using the docker image on `docker hub`: ```bash - docker pull opea/codegen:latest + docker pull opea/codegen:latest.1.0 ``` 2. Start services using the docker images built from source. See the [Gaudi Guide](./docker/gaudi/README.md) or [Xeon Guide](./docker/xeon/README.md) for more information. @@ -71,7 +71,7 @@ Find the corresponding [compose.yaml](./docker/gaudi/compose.yaml). ```bash cd GenAIExamples/CodeGen/docker/gaudi -docker compose up -d +TAG=0.1.0 docker compose up -d ``` > Notice: Currently only the **Habana Driver 1.16.x** is supported for Gaudi. @@ -84,7 +84,7 @@ Find the corresponding [compose.yaml](./docker/xeon/compose.yaml). ```bash cd GenAIExamples/CodeGen/docker/xeon -docker compose up -d +TAG=0.1.0 docker compose up -d ``` Refer to the [Xeon Guide](./docker/xeon/README.md) for more instructions on building docker images from source. diff --git a/CodeGen/codegen.yaml b/CodeGen/codegen.yaml index 7e4f423c33..d1ffef45fb 100644 --- a/CodeGen/codegen.yaml +++ b/CodeGen/codegen.yaml @@ -22,7 +22,7 @@ opea_micro_services: llm: host: ${LLM_SERVICE_HOST_IP} ports: ${LLM_SERVICE_PORT} - image: opea/llm-tgi:latest + image: opea/llm-tgi:latest.1.0 endpoint: /v1/chat/completions environment: TGI_LLM_ENDPOINT: ${TGI_LLM_ENDPOINT} @@ -31,7 +31,7 @@ opea_micro_services: host: ${UI_SERVICE_HOST_IP} ports: - "5173:5173" - image: opea/codegen-ui:latest + image: opea/codegen-ui:latest.1.0 environment: - CHAT_BASE_URL=${BACKEND_SERVICE_ENDPOINT} @@ -39,6 +39,6 @@ opea_mega_service: host: ${MEGA_SERVICE_HOST_IP} ports: ${MEGA_SERVICE_PORT} endpoint: /v1/codegen - image: opea/codegen:latest + image: opea/codegen:latest.1.0 mega_flow: - llm diff --git a/CodeGen/docker/gaudi/README.md b/CodeGen/docker/gaudi/README.md index a563f416ba..cf37e3793f 100644 --- a/CodeGen/docker/gaudi/README.md +++ b/CodeGen/docker/gaudi/README.md @@ -16,7 +16,7 @@ cd GenAIComps ### 2. Build the LLM Docker Image ```bash -docker build -t opea/llm-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . +docker build -t opea/llm-tgi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . ``` ### 3. Build the MegaService Docker Image @@ -26,7 +26,7 @@ To construct the Mega Service, we utilize the [GenAIComps](https://github.com/op ```bash git clone https://github.com/opea-project/GenAIExamples cd GenAIExamples/CodeGen/docker -docker build -t opea/codegen:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . +docker build -t opea/codegen:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . ``` ### 4. Build the UI Docker Image @@ -35,7 +35,7 @@ Construct the frontend Docker image via the command below: ```bash cd GenAIExamples/CodeGen/docker/ui/ -docker build -t opea/codegen-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . +docker build -t opea/codegen-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . ``` ### 5. Build the React UI Docker Image @@ -44,15 +44,15 @@ Construct the React frontend Docker image via the command below: ```bash cd GenAIExamples/CodeGen/docker/ui/ -docker build -t opea/codegen-react-ui:latest --build-arg BACKEND_SERVICE_ENDPOINT=$BACKEND_SERVICE_ENDPOINT -f ./docker/Dockerfile.react . +docker build -t opea/codegen-react-ui:latest.1.0 --build-arg BACKEND_SERVICE_ENDPOINT=$BACKEND_SERVICE_ENDPOINT -f ./docker/Dockerfile.react . ``` Then run the command `docker images`, you will have the following 3 Docker images: -- `opea/llm-tgi:latest` -- `opea/codegen:latest` -- `opea/codegen-ui:latest` -- `opea/codegen-react-ui:latest` +- `opea/llm-tgi:latest.1.0` +- `opea/codegen:latest.1.0` +- `opea/codegen-ui:latest.1.0` +- `opea/codegen-react-ui:latest.1.0` ## πŸš€ Start MicroServices and MegaService @@ -103,7 +103,7 @@ export BACKEND_SERVICE_ENDPOINT="http://${host_ip}:7778/v1/codegen" ```bash cd GenAIExamples/CodeGen/docker/gaudi -docker compose up -d +TAG=0.1.0 docker compose up -d ``` ### Validate the MicroServices and MegaService @@ -140,7 +140,7 @@ To access the frontend, open the following URL in your browser: `http://{host_ip ```yaml codegen-gaudi-ui-server: - image: opea/codegen-ui:latest + image: opea/codegen-ui:latest.1.0 ... ports: - "80:5173" diff --git a/CodeGen/docker/xeon/README.md b/CodeGen/docker/xeon/README.md index 74b5a2b7c2..c5dd7752df 100644 --- a/CodeGen/docker/xeon/README.md +++ b/CodeGen/docker/xeon/README.md @@ -24,7 +24,7 @@ cd GenAIComps ### 2. Build the LLM Docker Image ```bash -docker build -t opea/llm-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . +docker build -t opea/llm-tgi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . ``` ### 3. Build the MegaService Docker Image @@ -34,7 +34,7 @@ To construct the Mega Service, we utilize the [GenAIComps](https://github.com/op ```bash git clone https://github.com/opea-project/GenAIExamples cd GenAIExamples/CodeGen/docker -docker build -t opea/codegen:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . +docker build -t opea/codegen:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . ``` ### 4. Build the UI Docker Image @@ -43,14 +43,14 @@ Build the frontend Docker image via the command below: ```bash cd GenAIExamples/CodeGen/docker/ui/ -docker build -t opea/codegen-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . +docker build -t opea/codegen-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . ``` Then run the command `docker images`, you will have the following 3 Docker Images: -- `opea/llm-tgi:latest` -- `opea/codegen:latest` -- `opea/codegen-ui:latest` +- `opea/llm-tgi:latest.1.0` +- `opea/codegen:latest.1.0` +- `opea/codegen-ui:latest.1.0` ## πŸš€ Start Microservices and MegaService @@ -106,7 +106,7 @@ Note: Please replace the `host_ip` with you external IP address, do not use `loc ```bash cd GenAIExamples/CodeGen/docker/xeon -docker compose up -d +TAG=0.1.0 docker compose up -d ``` ### Validate the MicroServices and MegaService @@ -143,7 +143,7 @@ To access the frontend, open the following URL in your browser: `http://{host_ip ```yaml codegen-xeon-ui-server: - image: opea/codegen-ui:latest + image: opea/codegen-ui:latest.1.0 ... ports: - "80:5173" diff --git a/CodeGen/kubernetes/manifests/gaudi/codegen.yaml b/CodeGen/kubernetes/manifests/gaudi/codegen.yaml index 810c90f360..3ce0a65ff6 100644 --- a/CodeGen/kubernetes/manifests/gaudi/codegen.yaml +++ b/CodeGen/kubernetes/manifests/gaudi/codegen.yaml @@ -171,7 +171,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/llm-tgi:latest" + image: "opea/llm-tgi:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: llm-uservice @@ -296,7 +296,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/codegen:latest" + image: "opea/codegen:latest.1.0" imagePullPolicy: IfNotPresent volumeMounts: - mountPath: /tmp diff --git a/CodeGen/kubernetes/manifests/xeon/codegen.yaml b/CodeGen/kubernetes/manifests/xeon/codegen.yaml index 81f2b7a6e9..5dc8829df4 100644 --- a/CodeGen/kubernetes/manifests/xeon/codegen.yaml +++ b/CodeGen/kubernetes/manifests/xeon/codegen.yaml @@ -172,7 +172,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/llm-tgi:latest" + image: "opea/llm-tgi:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: llm-uservice @@ -296,7 +296,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/codegen:latest" + image: "opea/codegen:latest.1.0" imagePullPolicy: IfNotPresent volumeMounts: - mountPath: /tmp diff --git a/CodeGen/kubernetes/manifests/xeon/ui/react-codegen.yaml b/CodeGen/kubernetes/manifests/xeon/ui/react-codegen.yaml index 7ac1ab07ae..df09c77f11 100644 --- a/CodeGen/kubernetes/manifests/xeon/ui/react-codegen.yaml +++ b/CodeGen/kubernetes/manifests/xeon/ui/react-codegen.yaml @@ -179,7 +179,7 @@ spec: - name: no_proxy value: securityContext: {} - image: "opea/llm-tgi:latest" + image: "opea/llm-tgi:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: llm-uservice @@ -230,7 +230,7 @@ spec: - name: no_proxy value: securityContext: null - image: "opea/codegen:latest" + image: "opea/codegen:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: codegen @@ -273,7 +273,7 @@ spec: - name: no_proxy value: securityContext: null - image: "opea/codegen-react-ui:latest" + image: "opea/codegen-react-ui:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: react-ui diff --git a/CodeTrans/README.md b/CodeTrans/README.md index a56f02d641..102ae4352f 100644 --- a/CodeTrans/README.md +++ b/CodeTrans/README.md @@ -17,7 +17,7 @@ Currently we support two ways of deploying Code Translation services on docker: 1. Start services using the docker image on `docker hub`: ```bash - docker pull opea/codetrans:latest + docker pull opea/codetrans:latest.1.0 ``` 2. Start services using the docker images `built from source`: [Guide](./docker) @@ -57,7 +57,7 @@ Find the corresponding [compose.yaml](./docker/gaudi/compose.yaml). ```bash cd GenAIExamples/CodeTrans/docker/gaudi -docker compose up -d +TAG=0.1.0 docker compose up -d ``` > Notice: Currently only the **Habana Driver 1.16.x** is supported for Gaudi. @@ -70,7 +70,7 @@ Find the corresponding [compose.yaml](./docker/xeon/compose.yaml). ```bash cd GenAIExamples/CodeTrans/docker/xeon -docker compose up -d +TAG=0.1.0 docker compose up -d ``` Refer to the [Xeon Guide](./docker/xeon/README.md) for more instructions on building docker images from source. diff --git a/CodeTrans/codetrans.yaml b/CodeTrans/codetrans.yaml index a16c5a484c..2b80c15fa9 100644 --- a/CodeTrans/codetrans.yaml +++ b/CodeTrans/codetrans.yaml @@ -22,7 +22,7 @@ opea_micro_services: llm: host: ${LLM_SERVICE_HOST_IP} ports: ${LLM_SERVICE_PORT} - image: opea/llm-tgi:latest + image: opea/llm-tgi:latest.1.0 endpoint: /v1/chat/completions environment: TGI_LLM_ENDPOINT: ${TGI_LLM_ENDPOINT} @@ -31,7 +31,7 @@ opea_micro_services: host: ${UI_SERVICE_HOST_IP} ports: - "5173:5173" - image: opea/codetrans-ui:latest + image: opea/codetrans-ui:latest.1.0 environment: - CHAT_BASE_URL=${BACKEND_SERVICE_ENDPOINT} @@ -39,6 +39,6 @@ opea_mega_service: host: ${MEGA_SERVICE_HOST_IP} ports: ${MEGA_SERVICE_PORT} endpoint: /v1/codetrans - image: opea/codetrans:latest + image: opea/codetrans:latest.1.0 mega_flow: - llm diff --git a/CodeTrans/docker/gaudi/README.md b/CodeTrans/docker/gaudi/README.md index ab9c661ae3..4f730aace2 100755 --- a/CodeTrans/docker/gaudi/README.md +++ b/CodeTrans/docker/gaudi/README.md @@ -16,7 +16,7 @@ cd GenAIComps ### 2. Build the LLM Docker Image with the following command ```bash -docker build -t opea/llm-tgi:latest --no-cache --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . +docker build -t opea/llm-tgi:latest.1.0 --no-cache --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . ``` ### 3. Build MegaService Docker Image @@ -24,21 +24,21 @@ docker build -t opea/llm-tgi:latest --no-cache --build-arg https_proxy=$https_pr ```bash git clone https://github.com/opea-project/GenAIExamples.git cd GenAIExamples/CodeTrans/docker -docker build -t opea/codetrans:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . +docker build -t opea/codetrans:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . ``` ### 4. Build UI Docker Image ```bash cd GenAIExamples/CodeTrans/docker/ui -docker build -t opea/codetrans-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . +docker build -t opea/codetrans-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . ``` Then run the command `docker images`, you will have the following Docker Images: -- `opea/llm-tgi:latest` -- `opea/codetrans:latest` -- `opea/codetrans-ui:latest` +- `opea/llm-tgi:latest.1.0` +- `opea/codetrans:latest.1.0` +- `opea/codetrans-ui:latest.1.0` ## πŸš€ Start Microservices @@ -62,7 +62,7 @@ export BACKEND_SERVICE_ENDPOINT="http://${host_ip}:7777/v1/codetrans" ```bash cd GenAIExamples/CodeTrans/docker/gaudi -docker compose up -d +TAG=0.1.0 docker compose up -d ``` ### Validate Microservices diff --git a/CodeTrans/docker/xeon/README.md b/CodeTrans/docker/xeon/README.md index 270733f346..30d6a089fe 100755 --- a/CodeTrans/docker/xeon/README.md +++ b/CodeTrans/docker/xeon/README.md @@ -24,7 +24,7 @@ cd GenAIComps ### 2. Build the LLM Docker Image ```bash -docker build -t opea/llm-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . +docker build -t opea/llm-tgi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . ``` ### 3. Build MegaService Docker Image @@ -32,21 +32,21 @@ docker build -t opea/llm-tgi:latest --build-arg https_proxy=$https_proxy --build ```bash git clone https://github.com/opea-project/GenAIExamples.git cd GenAIExamples/CodeTrans/docker -docker build -t opea/codetrans:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . +docker build -t opea/codetrans:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . ``` ### 4. Build UI Docker Image ```bash cd GenAIExamples/CodeTrans/docker/ui -docker build -t opea/codetrans-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . +docker build -t opea/codetrans-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . ``` Then run the command `docker images`, you will have the following Docker Images: -- `opea/llm-tgi:latest` -- `opea/codetrans:latest` -- `opea/codetrans-ui:latest` +- `opea/llm-tgi:latest.1.0` +- `opea/codetrans:latest.1.0` +- `opea/codetrans-ui:latest.1.0` ## πŸš€ Start Microservices @@ -70,7 +70,7 @@ export BACKEND_SERVICE_ENDPOINT="http://${host_ip}:7777/v1/codetrans" ```bash cd GenAIExamples/CodeTrans/docker/xeon -docker compose up -d +TAG=0.1.0 docker compose up -d ``` ### Validate Microservices diff --git a/CodeTrans/kubernetes/manifests/gaudi/codetrans.yaml b/CodeTrans/kubernetes/manifests/gaudi/codetrans.yaml index b5b4ed1d9d..ee07286d9a 100644 --- a/CodeTrans/kubernetes/manifests/gaudi/codetrans.yaml +++ b/CodeTrans/kubernetes/manifests/gaudi/codetrans.yaml @@ -171,7 +171,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/llm-tgi:latest" + image: "opea/llm-tgi:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: llm-uservice @@ -296,7 +296,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/codetrans:latest" + image: "opea/codetrans:latest.1.0" imagePullPolicy: IfNotPresent volumeMounts: - mountPath: /tmp diff --git a/CodeTrans/kubernetes/manifests/xeon/codetrans.yaml b/CodeTrans/kubernetes/manifests/xeon/codetrans.yaml index a8cfdc8ff2..ef8b05102c 100644 --- a/CodeTrans/kubernetes/manifests/xeon/codetrans.yaml +++ b/CodeTrans/kubernetes/manifests/xeon/codetrans.yaml @@ -172,7 +172,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/llm-tgi:latest" + image: "opea/llm-tgi:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: llm-uservice @@ -296,7 +296,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/codetrans:latest" + image: "opea/codetrans:latest.1.0" imagePullPolicy: IfNotPresent volumeMounts: - mountPath: /tmp diff --git a/DocSum/README.md b/DocSum/README.md index 2372df8fc9..72219e05e8 100644 --- a/DocSum/README.md +++ b/DocSum/README.md @@ -20,7 +20,7 @@ Currently we support two ways of deploying Document Summarization services with 1. Start services using the docker image on `docker hub`: ```bash - docker pull opea/docsum:latest + docker pull opea/docsum:latest.1.0 ``` 2. Start services using the docker images `built from source`: [Guide](./docker) @@ -60,7 +60,7 @@ Find the corresponding [compose.yaml](./docker/gaudi/compose.yaml). ```bash cd GenAIExamples/DocSum/docker/gaudi/ -docker compose -f compose.yaml up -d +TAG=0.1.0 docker compose -f compose.yaml up -d ``` > Notice: Currently only the **Habana Driver 1.16.x** is supported for Gaudi. @@ -73,7 +73,7 @@ Find the corresponding [compose.yaml](./docker/xeon/compose.yaml). ```bash cd GenAIExamples/DocSum/docker/xeon/ -docker compose up -d +TAG=0.1.0 docker compose up -d ``` Refer to the [Xeon Guide](./docker/xeon/README.md) for more instructions on building docker images from source. diff --git a/DocSum/docker/gaudi/README.md b/DocSum/docker/gaudi/README.md index 742e7202fa..8d8e6c93d2 100644 --- a/DocSum/docker/gaudi/README.md +++ b/DocSum/docker/gaudi/README.md @@ -22,7 +22,7 @@ docker pull ghcr.io/huggingface/tgi-gaudi:2.0.1 ### 2. Build LLM Image ```bash -docker build -t opea/llm-docsum-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/summarization/tgi/Dockerfile . +docker build -t opea/llm-docsum-tgi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/summarization/tgi/Dockerfile . ``` ### 3. Build MegaService Docker Image @@ -32,7 +32,7 @@ To construct the Mega Service, we utilize the [GenAIComps](https://github.com/op ```bash git clone https://github.com/opea-project/GenAIExamples cd GenAIExamples/DocSum/docker -docker build -t opea/docsum:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . +docker build -t opea/docsum:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . ``` ### 4. Build UI Docker Image @@ -41,7 +41,7 @@ Construct the frontend Docker image using the command below: ```bash cd GenAIExamples/DocSum/docker/ui/ -docker build -t opea/docsum-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . +docker build -t opea/docsum-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . ``` ### 5. Build React UI Docker Image @@ -51,16 +51,16 @@ Build the frontend Docker image via below command: ```bash cd GenAIExamples/DocSum/docker/ui/ export BACKEND_SERVICE_ENDPOINT="http://${host_ip}:8888/v1/docsum" -docker build -t opea/docsum-react-ui:latest --build-arg BACKEND_SERVICE_ENDPOINT=$BACKEND_SERVICE_ENDPOINT -f ./docker/Dockerfile.react . +docker build -t opea/docsum-react-ui:latest.1.0 --build-arg BACKEND_SERVICE_ENDPOINT=$BACKEND_SERVICE_ENDPOINT -f ./docker/Dockerfile.react . ``` Then run the command `docker images`, you will have the following Docker Images: 1. `ghcr.io/huggingface/tgi-gaudi:2.0.1` -2. `opea/llm-docsum-tgi:latest` -3. `opea/docsum:latest` -4. `opea/docsum-ui:latest` -5. `opea/docsum-react-ui:latest` +2. `opea/llm-docsum-tgi:latest.1.0` +3. `opea/docsum:latest.1.0` +4. `opea/docsum-ui:latest.1.0` +5. `opea/docsum-react-ui:latest.1.0` ## πŸš€ Start Microservices and MegaService @@ -86,7 +86,7 @@ Note: Please replace with `host_ip` with your external IP address, do not use lo ```bash cd GenAIExamples/DocSum/docker/gaudi -docker compose up -d +TAG=0.1.0 docker compose up -d ``` ### Validate Microservices diff --git a/DocSum/docker/xeon/README.md b/DocSum/docker/xeon/README.md index 7d310c98b2..d401a1e895 100644 --- a/DocSum/docker/xeon/README.md +++ b/DocSum/docker/xeon/README.md @@ -22,7 +22,7 @@ cd GenAIComps ### 1. Build LLM Image ```bash -docker build -t opea/llm-docsum-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/summarization/tgi/Dockerfile . +docker build -t opea/llm-docsum-tgi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/summarization/tgi/Dockerfile . ``` Then run the command `docker images`, you will have the following four Docker Images: @@ -34,7 +34,7 @@ To construct the Mega Service, we utilize the [GenAIComps](https://github.com/op ```bash git clone https://github.com/opea-project/GenAIExamples cd GenAIExamples/DocSum/docker -docker build -t opea/docsum:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . +docker build -t opea/docsum:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . ``` ### 3. Build UI Docker Image @@ -43,14 +43,14 @@ Build the frontend Docker image via below command: ```bash cd GenAIExamples/DocSum/docker/ui/ -docker build -t opea/docsum-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . +docker build -t opea/docsum-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . ``` Then run the command `docker images`, you will have the following Docker Images: -1. `opea/llm-docsum-tgi:latest` -2. `opea/docsum:latest` -3. `opea/docsum-ui:latest` +1. `opea/llm-docsum-tgi:latest.1.0` +2. `opea/docsum:latest.1.0` +3. `opea/docsum-ui:latest.1.0` ### 4. Build React UI Docker Image @@ -59,15 +59,15 @@ Build the frontend Docker image via below command: ```bash cd GenAIExamples/DocSum/docker/ui/ export BACKEND_SERVICE_ENDPOINT="http://${host_ip}:8888/v1/docsum" -docker build -t opea/docsum-react-ui:latest --build-arg BACKEND_SERVICE_ENDPOINT=$BACKEND_SERVICE_ENDPOINT -f ./docker/Dockerfile.react . +docker build -t opea/docsum-react-ui:latest.1.0 --build-arg BACKEND_SERVICE_ENDPOINT=$BACKEND_SERVICE_ENDPOINT -f ./docker/Dockerfile.react . ``` Then run the command `docker images`, you will have the following Docker Images: -1. `opea/llm-docsum-tgi:latest` -2. `opea/docsum:latest` -3. `opea/docsum-ui:latest` -4. `opea/docsum-react-ui:latest` +1. `opea/llm-docsum-tgi:latest.1.0` +2. `opea/docsum:latest.1.0` +3. `opea/docsum-ui:latest.1.0` +4. `opea/docsum-react-ui:latest.1.0` ## πŸš€ Start Microservices and MegaService @@ -93,7 +93,7 @@ Note: Please replace with `host_ip` with your external IP address, do not use lo ```bash cd GenAIExamples/DocSum/docker/xeon -docker compose up -d +TAG=0.1.0 docker compose up -d ``` ### Validate Microservices diff --git a/DocSum/docsum.yaml b/DocSum/docsum.yaml index d4c8cb1c42..601e58a13a 100644 --- a/DocSum/docsum.yaml +++ b/DocSum/docsum.yaml @@ -22,7 +22,7 @@ opea_micro_services: llm: host: ${LLM_SERVICE_HOST_IP} ports: ${LLM_SERVICE_PORT} - image: opea/llm-tgi:latest + image: opea/llm-tgi:latest.1.0 endpoint: /v1/chat/completions environment: TGI_LLM_ENDPOINT: ${TGI_LLM_ENDPOINT} @@ -31,14 +31,14 @@ opea_micro_services: host: ${UI_SERVICE_HOST_IP} ports: - "5173:5173" - image: opea/docsum-ui:latest + image: opea/docsum-ui:latest.1.0 environment: - CHAT_BASE_URL=${BACKEND_SERVICE_ENDPOINT} opea_mega_service: host: ${MEGA_SERVICE_HOST_IP} ports: ${MEGA_SERVICE_PORT} - image: opea/docsum:latest + image: opea/docsum:latest.1.0 endpoint: /v1/docsum mega_flow: - llm diff --git a/DocSum/kubernetes/manifests/gaudi/docsum.yaml b/DocSum/kubernetes/manifests/gaudi/docsum.yaml index 9eac0d0627..a7f852403b 100644 --- a/DocSum/kubernetes/manifests/gaudi/docsum.yaml +++ b/DocSum/kubernetes/manifests/gaudi/docsum.yaml @@ -171,7 +171,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/llm-docsum-tgi:latest" + image: "opea/llm-docsum-tgi:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: llm-uservice @@ -296,7 +296,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/docsum:latest" + image: "opea/docsum:latest.1.0" imagePullPolicy: IfNotPresent volumeMounts: - mountPath: /tmp diff --git a/DocSum/kubernetes/manifests/xeon/docsum.yaml b/DocSum/kubernetes/manifests/xeon/docsum.yaml index 3cfb3c682e..131615335f 100644 --- a/DocSum/kubernetes/manifests/xeon/docsum.yaml +++ b/DocSum/kubernetes/manifests/xeon/docsum.yaml @@ -172,7 +172,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/llm-docsum-tgi:latest" + image: "opea/llm-docsum-tgi:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: llm-uservice @@ -296,7 +296,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/docsum:latest" + image: "opea/docsum:latest.1.0" imagePullPolicy: IfNotPresent volumeMounts: - mountPath: /tmp diff --git a/DocSum/kubernetes/manifests/xeon/ui/react-docsum.yaml b/DocSum/kubernetes/manifests/xeon/ui/react-docsum.yaml index 395059e02b..1e13d86752 100644 --- a/DocSum/kubernetes/manifests/xeon/ui/react-docsum.yaml +++ b/DocSum/kubernetes/manifests/xeon/ui/react-docsum.yaml @@ -180,7 +180,7 @@ spec: value: securityContext: {} - image: "opea/llm-docsum-tgi:latest" + image: "opea/llm-docsum-tgi:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: llm-uservice @@ -231,7 +231,7 @@ spec: - name: no_proxy value: securityContext: null - image: "opea/docsum:latest" + image: "opea/docsum:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: docsum @@ -274,7 +274,7 @@ spec: - name: no_proxy value: securityContext: null - image: "opea/docsum-react-ui:latest" + image: "opea/docsum-react-ui:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: react-ui diff --git a/FaqGen/docker/gaudi/README.md b/FaqGen/docker/gaudi/README.md index 509cbface7..c8054f7936 100644 --- a/FaqGen/docker/gaudi/README.md +++ b/FaqGen/docker/gaudi/README.md @@ -22,7 +22,7 @@ docker pull ghcr.io/huggingface/tgi-gaudi:2.0.1 ### 2. Build LLM Image ```bash -docker build -t opea/llm-faqgen-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/faq-generation/tgi/Dockerfile . +docker build -t opea/llm-faqgen-tgi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/faq-generation/tgi/Dockerfile . ``` ### 3. Build MegaService Docker Image @@ -32,7 +32,7 @@ To construct the Mega Service, we utilize the [GenAIComps](https://github.com/op ```bash git clone https://github.com/opea-project/GenAIExamples cd GenAIExamples/FaqGen/docker/ -docker build --no-cache -t opea/faqgen:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . +docker build --no-cache -t opea/faqgen:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . ``` ### 4. Build UI Docker Image @@ -41,7 +41,7 @@ Construct the frontend Docker image using the command below: ```bash cd GenAIExamples/FaqGen/docker/ui/ -docker build -t opea/faqgen-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . +docker build -t opea/faqgen-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . ``` ### 5. Build react UI Docker Image (Optional) @@ -51,16 +51,16 @@ Build the frontend Docker image based on react framework via below command: ```bash cd GenAIExamples/FaqGen/docker/ui export BACKEND_SERVICE_ENDPOINT="http://${host_ip}:8888/v1/faqgen" -docker build -t opea/faqgen-react-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy --build-arg BACKEND_SERVICE_ENDPOINT=$BACKEND_SERVICE_ENDPOINT -f ./docker/Dockerfile.react . +docker build -t opea/faqgen-react-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy --build-arg BACKEND_SERVICE_ENDPOINT=$BACKEND_SERVICE_ENDPOINT -f ./docker/Dockerfile.react . ``` Then run the command `docker images`, you will have the following Docker Images: 1. `ghcr.io/huggingface/tgi-gaudi:2.0.1` -2. `opea/llm-faqgen-tgi:latest` -3. `opea/faqgen:latest` -4. `opea/faqgen-ui:latest` -5. `opea/faqgen-react-ui:latest` +2. `opea/llm-faqgen-tgi:latest.1.0` +3. `opea/faqgen:latest.1.0` +4. `opea/faqgen-ui:latest.1.0` +5. `opea/faqgen-react-ui:latest.1.0` ## πŸš€ Start Microservices and MegaService @@ -86,7 +86,7 @@ Note: Please replace with `host_ip` with your external IP address, do not use lo ```bash cd GenAIExamples/FaqGen/docker/gaudi -docker compose up -d +TAG=0.1.0 docker compose up -d ``` ### Validate Microservices @@ -129,7 +129,7 @@ To access the FAQGen (react based) frontend, modify the UI service in the `compo ```bash faqgen-xeon-react-ui-server: - image: opea/faqgen-react-ui:latest + image: opea/faqgen-react-ui:latest.1.0 container_name: faqgen-xeon-react-ui-server environment: - no_proxy=${no_proxy} diff --git a/FaqGen/docker/xeon/README.md b/FaqGen/docker/xeon/README.md index e86e6480bc..ff8ef170e6 100644 --- a/FaqGen/docker/xeon/README.md +++ b/FaqGen/docker/xeon/README.md @@ -22,7 +22,7 @@ cd GenAIComps ### 1. Build LLM Image ```bash -docker build -t opea/llm-faqgen-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/faq-generation/tgi/Dockerfile . +docker build -t opea/llm-faqgen-tgi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/faq-generation/tgi/Dockerfile . ``` ### 2. Build MegaService Docker Image @@ -32,7 +32,7 @@ To construct the Mega Service, we utilize the [GenAIComps](https://github.com/op ```bash git clone https://github.com/opea-project/GenAIExamples cd GenAIExamples/FaqGen/docker/ -docker build --no-cache -t opea/faqgen:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f GenAIExamples/FaqGen/docker/Dockerfile . +docker build --no-cache -t opea/faqgen:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f GenAIExamples/FaqGen/docker/Dockerfile . ``` ### 3. Build UI Docker Image @@ -41,7 +41,7 @@ Build the frontend Docker image via below command: ```bash cd GenAIExamples/FaqGen/docker/ui/ -docker build -t opea/faqgen-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . +docker build -t opea/faqgen-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . ``` ### 4. Build react UI Docker Image (Optional) @@ -51,15 +51,15 @@ Build the frontend Docker image based on react framework via below command: ```bash cd GenAIExamples/FaqGen/docker/ui export BACKEND_SERVICE_ENDPOINT="http://${host_ip}:8888/v1/faqgen" -docker build -t opea/faqgen-react-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy --build-arg BACKEND_SERVICE_ENDPOINT=$BACKEND_SERVICE_ENDPOINT -f ./docker/Dockerfile.react . +docker build -t opea/faqgen-react-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy --build-arg BACKEND_SERVICE_ENDPOINT=$BACKEND_SERVICE_ENDPOINT -f ./docker/Dockerfile.react . ``` Then run the command `docker images`, you will have the following Docker Images: -1. `opea/llm-faqgen-tgi:latest` -2. `opea/faqgen:latest` -3. `opea/faqgen-ui:latest` -4. `opea/faqgen-react-ui:latest` +1. `opea/llm-faqgen-tgi:latest.1.0` +2. `opea/faqgen:latest.1.0` +3. `opea/faqgen-ui:latest.1.0` +4. `opea/faqgen-react-ui:latest.1.0` ## πŸš€ Start Microservices and MegaService @@ -85,7 +85,7 @@ Note: Please replace with `host_ip` with your external IP address, do not use lo ```bash cd GenAIExamples/FaqGen/docker/xeon -docker compose up -d +TAG=0.1.0 docker compose up -d ``` ### Validate Microservices @@ -130,7 +130,7 @@ To access the FAQGen (react based) frontend, modify the UI service in the `compo ```bash faqgen-xeon-react-ui-server: - image: opea/faqgen-react-ui:latest + image: opea/faqgen-react-ui:latest.1.0 container_name: faqgen-xeon-react-ui-server environment: - no_proxy=${no_proxy} diff --git a/FaqGen/faqgen.yaml b/FaqGen/faqgen.yaml index 8832c0c939..a31877d384 100644 --- a/FaqGen/faqgen.yaml +++ b/FaqGen/faqgen.yaml @@ -22,7 +22,7 @@ opea_micro_services: llm: host: ${LLM_SERVICE_HOST_IP} ports: ${LLM_SERVICE_PORT} - image: opea/llm-tgi:latest + image: opea/llm-tgi:latest.1.0 endpoint: /v1/chat/completions environment: TGI_LLM_ENDPOINT: ${TGI_LLM_ENDPOINT} @@ -31,14 +31,14 @@ opea_micro_services: host: ${UI_SERVICE_HOST_IP} ports: - "5173:5173" - image: opea/faqgen-ui:latest + image: opea/faqgen-ui:latest.1.0 environment: - CHAT_BASE_URL=${BACKEND_SERVICE_ENDPOINT} opea_mega_service: host: ${MEGA_SERVICE_HOST_IP} ports: ${MEGA_SERVICE_PORT} - image: opea/faqgen:latest + image: opea/faqgen:latest.1.0 endpoint: /v1/faqgen mega_flow: - llm diff --git a/FaqGen/kubernetes/manifests/gaudi/faqgen.yaml b/FaqGen/kubernetes/manifests/gaudi/faqgen.yaml index 76a68080c1..c00e6f6481 100644 --- a/FaqGen/kubernetes/manifests/gaudi/faqgen.yaml +++ b/FaqGen/kubernetes/manifests/gaudi/faqgen.yaml @@ -117,7 +117,7 @@ spec: value: "http://faq-tgi-svc.default.svc.cluster.local:8010" - name: HUGGINGFACEHUB_API_TOKEN value: "insert-your-huggingface-token-here" - image: opea/llm-faqgen-tgi:latest + image: opea/llm-faqgen-tgi:latest.1.0 imagePullPolicy: IfNotPresent args: null ports: @@ -166,7 +166,7 @@ spec: value: faq-mega-server-svc - name: MEGA_SERVICE_PORT value: "7777" - image: opea/faqgen:latest + image: opea/faqgen:latest.1.0 imagePullPolicy: IfNotPresent args: null ports: diff --git a/FaqGen/kubernetes/manifests/ui.yaml b/FaqGen/kubernetes/manifests/ui.yaml index f74299a094..e3916ff13c 100644 --- a/FaqGen/kubernetes/manifests/ui.yaml +++ b/FaqGen/kubernetes/manifests/ui.yaml @@ -24,7 +24,7 @@ spec: env: - name: DOC_BASE_URL value: http://{insert_your_ip_here}:7779/v1/faqgen - image: opea/faqgen-ui:latest + image: opea/faqgen-ui:latest.1.0 imagePullPolicy: IfNotPresent args: null ports: diff --git a/FaqGen/kubernetes/manifests/xeon/faqgen.yaml b/FaqGen/kubernetes/manifests/xeon/faqgen.yaml index ddf81afdf4..5e3bd1e097 100644 --- a/FaqGen/kubernetes/manifests/xeon/faqgen.yaml +++ b/FaqGen/kubernetes/manifests/xeon/faqgen.yaml @@ -96,7 +96,7 @@ spec: value: "http://faq-tgi-cpu-svc.default.svc.cluster.local:8011" - name: HUGGINGFACEHUB_API_TOKEN value: "insert-your-huggingface-token-here" - image: opea/llm-faqgen-tgi:latest + image: opea/llm-faqgen-tgi:latest.1.0 imagePullPolicy: IfNotPresent args: null ports: @@ -145,7 +145,7 @@ spec: value: faq-mega-server-cpu-svc - name: MEGA_SERVICE_PORT value: "7777" - image: opea/faqgen:latest + image: opea/faqgen:latest.1.0 imagePullPolicy: IfNotPresent args: null ports: diff --git a/FaqGen/kubernetes/manifests/xeon/ui/react-faqgen.yaml b/FaqGen/kubernetes/manifests/xeon/ui/react-faqgen.yaml index f4ae09d2ca..06aae2bf03 100644 --- a/FaqGen/kubernetes/manifests/xeon/ui/react-faqgen.yaml +++ b/FaqGen/kubernetes/manifests/xeon/ui/react-faqgen.yaml @@ -179,7 +179,7 @@ spec: - name: no_proxy value: securityContext: {} - image: "opea/llm-faqgen-tgi:latest" + image: "opea/llm-faqgen-tgi:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: llm-uservice @@ -230,7 +230,7 @@ spec: - name: no_proxy value: securityContext: null - image: "opea/faqgen:latest" + image: "opea/faqgen:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: faqgen @@ -273,7 +273,7 @@ spec: - name: no_proxy value: securityContext: null - image: "opea/faqgen-react-ui:latest" + image: "opea/faqgen-react-ui:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: react-ui diff --git a/SearchQnA/README.md b/SearchQnA/README.md index 1b83f664d4..e908ca7ac7 100644 --- a/SearchQnA/README.md +++ b/SearchQnA/README.md @@ -29,7 +29,7 @@ Currently we support two ways of deploying SearchQnA services with docker compos 1. Start services using the docker image on `docker hub`: ```bash - docker pull opea/searchqna:latest + docker pull opea/searchqna:latest.1.0 ``` 2. Start services using the docker images `built from source`: [Guide](./docker) @@ -69,7 +69,7 @@ If your version of `Habana Driver` < 1.16.0 (check with `hl-smi`), run the follo ```bash cd GenAIExamples/SearchQnA/docker/gaudi/ -docker compose up -d +TAG=0.1.0 docker compose up -d ``` > Notice: Currently only the **Habana Driver 1.16.x** is supported for Gaudi. @@ -82,7 +82,7 @@ Find the corresponding [compose.yaml](./docker/xeon/compose.yaml). ```bash cd GenAIExamples/SearchQnA/docker/xeon/ -docker compose up -d +TAG=0.1.0 docker compose up -d ``` Refer to the [Xeon Guide](./docker/xeon/README.md) for more instructions on building docker images from source. diff --git a/SearchQnA/docker/gaudi/README.md b/SearchQnA/docker/gaudi/README.md index 765ec16183..f6ee3823d0 100644 --- a/SearchQnA/docker/gaudi/README.md +++ b/SearchQnA/docker/gaudi/README.md @@ -16,25 +16,25 @@ cd GenAIComps ### 2. Build Embedding Image ```bash -docker build --no-cache -t opea/embedding-tei:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/embeddings/langchain/docker/Dockerfile . +docker build --no-cache -t opea/embedding-tei:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/embeddings/langchain/docker/Dockerfile . ``` ### 3. Build Retriever Image ```bash -docker build --no-cache -t opea/web-retriever-chroma:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/web_retrievers/langchain/chroma/docker/Dockerfile . +docker build --no-cache -t opea/web-retriever-chroma:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/web_retrievers/langchain/chroma/docker/Dockerfile . ``` ### 4. Build Rerank Image ```bash -docker build --no-cache -t opea/reranking-tei:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/reranks/tei/docker/Dockerfile . +docker build --no-cache -t opea/reranking-tei:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/reranks/tei/docker/Dockerfile . ``` ### 5. Build LLM Image ```bash -docker build --no-cache -t opea/llm-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . +docker build --no-cache -t opea/llm-tgi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . ``` ### 6. Build TEI Gaudi Image @@ -44,7 +44,7 @@ Since a TEI Gaudi Docker image hasn't been published, we'll need to build it fro ```bash git clone https://github.com/huggingface/tei-gaudi cd tei-gaudi/ -docker build --no-cache -f Dockerfile-hpu -t opea/tei-gaudi:latest . +docker build --no-cache -f Dockerfile-hpu -t opea/tei-gaudi:latest.1.0 . cd ../.. ``` @@ -55,25 +55,25 @@ To construct the Mega Service, we utilize the [GenAIComps](https://github.com/op ```bash git clone https://github.com/opea-project/GenAIExamples.git cd GenAIExamples/SearchQnA/docker -docker build --no-cache -t opea/searchqna:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . +docker build --no-cache -t opea/searchqna:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . cd ../../.. ``` -Then you need to build the last Docker image `opea/searchqna:latest`, which represents the Mega service through following commands: +Then you need to build the last Docker image `opea/searchqna:latest.1.0`, which represents the Mega service through following commands: ```bash cd GenAIExamples/SearchQnA/docker -docker build --no-cache -t opea/searchqna:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . +docker build --no-cache -t opea/searchqna:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . ``` Then run the command `docker images`, you will have -1. `opea/tei-gaudi:latest` -2. `opea/embedding-tei:latest` -3. `opea/web-retriever-chroma:latest` -4. `opea/reranking-tei:latest` -5. `opea/llm-tgi:latest` -6. `opea/searchqna:latest` +1. `opea/tei-gaudi:latest.1.0` +2. `opea/embedding-tei:latest.1.0` +3. `opea/web-retriever-chroma:latest.1.0` +4. `opea/reranking-tei:latest.1.0` +5. `opea/llm-tgi:latest.1.0` +6. `opea/searchqna:latest.1.0` ## πŸš€ Set the environment variables @@ -109,7 +109,7 @@ export LLM_SERVICE_PORT=3007 ```bash cd GenAIExamples/SearchQnA/docker/gaudi/ -docker compose up -d +TAG=0.1.0 docker compose up -d ``` ## πŸš€ Test MicroServices diff --git a/SearchQnA/docker/xeon/README.md b/SearchQnA/docker/xeon/README.md index 54068905d2..0b8862f9b4 100644 --- a/SearchQnA/docker/xeon/README.md +++ b/SearchQnA/docker/xeon/README.md @@ -14,25 +14,25 @@ cd GenAIComps ### 2. Build Embedding Image ```bash -docker build --no-cache -t opea/embedding-tei:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/embeddings/langchain/docker/Dockerfile . +docker build --no-cache -t opea/embedding-tei:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/embeddings/langchain/docker/Dockerfile . ``` ### 3. Build Retriever Image ```bash -docker build --no-cache -t opea/web-retriever-chroma:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/web_retrievers/langchain/chroma/docker/Dockerfile . +docker build --no-cache -t opea/web-retriever-chroma:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/web_retrievers/langchain/chroma/docker/Dockerfile . ``` ### 4. Build Rerank Image ```bash -docker build --no-cache -t opea/reranking-tei:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/reranks/tei/docker/Dockerfile . +docker build --no-cache -t opea/reranking-tei:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/reranks/tei/docker/Dockerfile . ``` ### 5. Build LLM Image ```bash -docker build --no-cache -t opea/llm-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . +docker build --no-cache -t opea/llm-tgi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . ``` ### 6. Build MegaService Docker Image @@ -42,17 +42,17 @@ To construct the Mega Service, we utilize the [GenAIComps](https://github.com/op ```bash git clone https://github.com/opea-project/GenAIExamples.git cd GenAIExamples/SearchQnA/docker -docker build --no-cache -t opea/searchqna:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . +docker build --no-cache -t opea/searchqna:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . cd ../../.. ``` Then run the command `docker images`, you will have following images ready: -1. `opea/embedding-tei:latest` -2. `opea/web-retriever-chroma:latest` -3. `opea/reranking-tei:latest` -4. `opea/llm-tgi:latest` -5. `opea/searchqna:latest` +1. `opea/embedding-tei:latest.1.0` +2. `opea/web-retriever-chroma:latest.1.0` +3. `opea/reranking-tei:latest.1.0` +4. `opea/llm-tgi:latest.1.0` +5. `opea/searchqna:latest.1.0` ## πŸš€ Set the environment variables @@ -88,7 +88,7 @@ export LLM_SERVICE_PORT=3007 ```bash cd GenAIExamples/SearchQnA/docker/xeon/ -docker compose up -d +TAG=0.1.0 docker compose up -d ``` ## πŸš€ Test MicroServices diff --git a/Translation/docker/gaudi/README.md b/Translation/docker/gaudi/README.md index 04ec63453e..cc3bccc64c 100644 --- a/Translation/docker/gaudi/README.md +++ b/Translation/docker/gaudi/README.md @@ -14,7 +14,7 @@ cd GenAIComps ### 1. Build LLM Image ```bash -docker build -t opea/llm-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . +docker build -t opea/llm-tgi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . ``` ### 2. Build MegaService Docker Image @@ -24,7 +24,7 @@ To construct the Mega Service, we utilize the [GenAIComps](https://github.com/op ```bash git clone https://github.com/opea-project/GenAIExamples cd GenAIExamples/Translation/docker -docker build -t opea/translation:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . +docker build -t opea/translation:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . ``` ### 3. Build UI Docker Image @@ -33,14 +33,14 @@ Construct the frontend Docker image using the command below: ```bash cd GenAIExamples/Translation/docker/ui/ -docker build -t opea/translation-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . +docker build -t opea/translation-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . ``` Then run the command `docker images`, you will have the following four Docker Images: -1. `opea/llm-tgi:latest` -2. `opea/translation:latest` -3. `opea/translation-ui:latest` +1. `opea/llm-tgi:latest.1.0` +2. `opea/translation:latest.1.0` +3. `opea/translation-ui:latest.1.0` ## πŸš€ Start Microservices @@ -64,7 +64,7 @@ Note: Please replace with `host_ip` with you external IP address, do not use loc ### Start Microservice Docker Containers ```bash -docker compose up -d +TAG=0.1.0 docker compose up -d ``` ### Validate Microservices diff --git a/Translation/docker/xeon/README.md b/Translation/docker/xeon/README.md index 915e2b3a4e..1c73ff3e95 100644 --- a/Translation/docker/xeon/README.md +++ b/Translation/docker/xeon/README.md @@ -22,7 +22,7 @@ cd GenAIComps ### 1. Build LLM Image ```bash -docker build -t opea/llm-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . +docker build -t opea/llm-tgi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . ``` ### 2. Build MegaService Docker Image @@ -32,7 +32,7 @@ To construct the Mega Service, we utilize the [GenAIComps](https://github.com/op ```bash git clone https://github.com/opea-project/GenAIExamples cd GenAIExamples/Translation/docker -docker build -t opea/translation:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . +docker build -t opea/translation:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . ``` ### 3. Build UI Docker Image @@ -41,14 +41,14 @@ Build frontend Docker image via below command: ```bash cd GenAIExamples/Translation/docker/ui -docker build -t opea/translation-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . +docker build -t opea/translation-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . ``` Then run the command `docker images`, you will have the following Docker Images: -1. `opea/llm-tgi:latest` -2. `opea/translation:latest` -3. `opea/translation-ui:latest` +1. `opea/llm-tgi:latest.1.0` +2. `opea/translation:latest.1.0` +3. `opea/translation-ui:latest.1.0` ## πŸš€ Start Microservices @@ -72,7 +72,7 @@ Note: Please replace with `host_ip` with you external IP address, do not use loc ### Start Microservice Docker Containers ```bash -docker compose up -d +TAG=0.1.0 docker compose up -d ``` ### Validate Microservices diff --git a/Translation/translation.yaml b/Translation/translation.yaml index 23aac46758..fff9264b06 100644 --- a/Translation/translation.yaml +++ b/Translation/translation.yaml @@ -22,7 +22,7 @@ opea_micro_services: llm: host: ${LLM_SERVICE_HOST_IP} ports: ${LLM_SERVICE_PORT} - image: opea/llm-tgi:latest + image: opea/llm-tgi:latest.1.0 endpoint: /v1/chat/completions environment: TGI_LLM_ENDPOINT: ${TGI_LLM_ENDPOINT} @@ -31,14 +31,14 @@ opea_micro_services: host: ${UI_SERVICE_HOST_IP} ports: - "5173:5173" - image: opea/translation-ui:latest + image: opea/translation-ui:latest.1.0 environment: - CHAT_BASE_URL=${BACKEND_SERVICE_ENDPOINT} opea_mega_service: host: ${MEGA_SERVICE_HOST_IP} ports: ${MEGA_SERVICE_PORT} - image: opea/translation:latest + image: opea/translation:latest.1.0 endpoint: /v1/translation mega_flow: - llm diff --git a/VisualQnA/README.md b/VisualQnA/README.md index 757a2657c8..a03c6ee198 100644 --- a/VisualQnA/README.md +++ b/VisualQnA/README.md @@ -63,7 +63,7 @@ Find the corresponding [compose.yaml](./docker/gaudi/compose.yaml). ```bash cd GenAIExamples/VisualQnA/docker/gaudi/ -docker compose up -d +TAG=0.1.0 docker compose up -d ``` > Notice: Currently only the **Habana Driver 1.16.x** is supported for Gaudi. @@ -76,5 +76,5 @@ Find the corresponding [compose.yaml](./docker/xeon/compose.yaml). ```bash cd GenAIExamples/VisualQnA/docker/xeon/ -docker compose up -d +TAG=0.1.0 docker compose up -d ``` diff --git a/VisualQnA/docker/gaudi/README.md b/VisualQnA/docker/gaudi/README.md index 72d5efb07f..eddab34738 100644 --- a/VisualQnA/docker/gaudi/README.md +++ b/VisualQnA/docker/gaudi/README.md @@ -16,7 +16,7 @@ cd GenAIComps ### 2. Build LLM Image ```bash -docker build --no-cache -t opea/lvm-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/lvms/Dockerfile_tgi . +docker build --no-cache -t opea/lvm-tgi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/lvms/Dockerfile_tgi . ``` ### 3. Build TGI Gaudi Image @@ -27,7 +27,7 @@ Since TGI Gaudi has not supported llava-next in main branch, we'll need to build git clone https://github.com/yuanwu2017/tgi-gaudi.git cd tgi-gaudi/ git checkout v2.0.4 -docker build -t opea/llava-tgi:latest . +docker build -t opea/llava-tgi:latest.1.0 . cd ../ ``` @@ -38,7 +38,7 @@ To construct the Mega Service, we utilize the [GenAIComps](https://github.com/op ```bash git clone https://github.com/opea-project/GenAIExamples.git cd GenAIExamples/VisualQnA/docker -docker build --no-cache -t opea/visualqna:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . +docker build --no-cache -t opea/visualqna:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . cd ../../.. ``` @@ -48,16 +48,16 @@ Build frontend Docker image via below command: ```bash cd GenAIExamples/VisualQnA/docker/ui/ -docker build --no-cache -t opea/visualqna-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . +docker build --no-cache -t opea/visualqna-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . cd ../../../.. ``` Then run the command `docker images`, you will have the following 4 Docker Images: -1. `opea/llava-tgi:latest` -2. `opea/lvm-tgi:latest` -3. `opea/visualqna:latest` -4. `opea/visualqna-ui:latest` +1. `opea/llava-tgi:latest.1.0` +2. `opea/lvm-tgi:latest.1.0` +3. `opea/visualqna:latest.1.0` +4. `opea/visualqna-ui:latest.1.0` ## πŸš€ Start MicroServices and MegaService @@ -86,7 +86,7 @@ cd GenAIExamples/VisualQnA/docker/gaudi/ ``` ```bash -docker compose -f compose.yaml up -d +TAG=0.1.0 docker compose -f compose.yaml up -d ``` > **_NOTE:_** Users need at least one Gaudi cards to run the VisualQnA successfully. @@ -132,7 +132,7 @@ To access the frontend, open the following URL in your browser: http://{host_ip} ```yaml visualqna-gaudi-ui-server: - image: opea/visualqna-ui:latest + image: opea/visualqna-ui:latest.1.0 ... ports: - "80:5173" diff --git a/VisualQnA/docker/xeon/README.md b/VisualQnA/docker/xeon/README.md index 8f3896c056..82220973a4 100644 --- a/VisualQnA/docker/xeon/README.md +++ b/VisualQnA/docker/xeon/README.md @@ -44,7 +44,7 @@ cd GenAIComps ### 1. Build LVM Image ```bash -docker build --no-cache -t opea/lvm-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/lvms/Dockerfile_tgi . +docker build --no-cache -t opea/lvm-tgi:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/lvms/Dockerfile_tgi . ``` ### 2. Build MegaService Docker Image @@ -54,7 +54,7 @@ To construct the Mega Service, we utilize the [GenAIComps](https://github.com/op ```bash git clone https://github.com/opea-project/GenAIExamples.git cd GenAIExamples/VisualQnA/docker -docker build --no-cache -t opea/visualqna:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . +docker build --no-cache -t opea/visualqna:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . cd ../../.. ``` @@ -64,7 +64,7 @@ Build frontend Docker image via below command: ```bash cd GenAIExamples/VisualQnA/docker/ui/ -docker build --no-cache -t opea/visualqna-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . +docker build --no-cache -t opea/visualqna-ui:latest.1.0 --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . cd ../../../.. ``` @@ -77,9 +77,9 @@ docker pull ghcr.io/huggingface/text-generation-inference:2.2.0 Then run the command `docker images`, you will have the following 4 Docker Images: 1. `ghcr.io/huggingface/text-generation-inference:2.2.0` -2. `opea/lvm-tgi:latest` -3. `opea/visualqna:latest` -4. `opea/visualqna-ui:latest` +2. `opea/lvm-tgi:latest.1.0` +3. `opea/visualqna:latest.1.0` +4. `opea/visualqna-ui:latest.1.0` ## πŸš€ Start Microservices @@ -124,7 +124,7 @@ cd GenAIExamples/VisualQnA/docker/xeon/ ``` ```bash -docker compose -f compose.yaml up -d +TAG=0.1.0 docker compose -f compose.yaml up -d ``` ### Validate Microservices @@ -168,7 +168,7 @@ To access the frontend, open the following URL in your browser: http://{host_ip} ```yaml visualqna-gaudi-ui-server: - image: opea/visualqna-ui:latest + image: opea/visualqna-ui:latest.1.0 ... ports: - "80:5173" diff --git a/VisualQnA/kubernetes/manifests/gaudi/visualqna.yaml b/VisualQnA/kubernetes/manifests/gaudi/visualqna.yaml index 6f00bfb8ce..85c949153c 100644 --- a/VisualQnA/kubernetes/manifests/gaudi/visualqna.yaml +++ b/VisualQnA/kubernetes/manifests/gaudi/visualqna.yaml @@ -165,7 +165,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/lvm-tgi:latest" + image: "opea/lvm-tgi:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: lvm-uservice @@ -215,7 +215,7 @@ spec: name: visualqna-tgi-config securityContext: {} - image: "opea/llava-tgi:latest" + image: "opea/llava-tgi:latest.1.0" imagePullPolicy: IfNotPresent volumeMounts: - mountPath: /data @@ -282,7 +282,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/visualqna:latest" + image: "opea/visualqna:latest.1.0" imagePullPolicy: IfNotPresent volumeMounts: - mountPath: /tmp diff --git a/VisualQnA/kubernetes/manifests/xeon/visualqna.yaml b/VisualQnA/kubernetes/manifests/xeon/visualqna.yaml index 744853d810..e3b6b4ec7f 100644 --- a/VisualQnA/kubernetes/manifests/xeon/visualqna.yaml +++ b/VisualQnA/kubernetes/manifests/xeon/visualqna.yaml @@ -166,7 +166,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/lvm-tgi:latest" + image: "opea/lvm-tgi:latest.1.0" imagePullPolicy: IfNotPresent ports: - name: lvm-uservice @@ -282,7 +282,7 @@ spec: runAsUser: 1000 seccompProfile: type: RuntimeDefault - image: "opea/visualqna:latest" + image: "opea/visualqna:latest.1.0" imagePullPolicy: IfNotPresent volumeMounts: - mountPath: /tmp