Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

GMC: adopt new common/menifests #203

Merged
merged 3 commits into from
Jul 22, 2024
Merged
Show file tree
Hide file tree
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 3 additions & 1 deletion .github/workflows/scripts/e2e/gmc_install.sh
Original file line number Diff line number Diff line change
Expand Up @@ -36,6 +36,7 @@ function copy_manifests() {
# Copy manifest into gmc
mkdir -p $(pwd)/config/manifests
cp $(dirname $(pwd))/manifests/ChatQnA/*.yaml -p $(pwd)/config/manifests/
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

remove coping from ChatQnA after all manifests finalized

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

yes, that is planned

cp $(dirname $(pwd))/manifests/common/*.yaml -p $(pwd)/config/manifests/
cp $(dirname $(pwd))/manifests/DocSum/xeon/docsum_llm.yaml -p $(pwd)/config/manifests/
cp $(dirname $(pwd))/manifests/DocSum/gaudi/docsum_gaudi_llm.yaml -p $(pwd)/config/manifests/
}
Expand All @@ -59,7 +60,8 @@ function init_gmc() {
sed -i "s|name: system|name: $SYSTEM_NAMESPACE|g" $(pwd)/config/rbac/gmc-manager-rbac.yaml

# replace the mount dir "path: /mnt/model" with "path: $CHART_MOUNT"
find . -name '*.yaml' -type f -exec sed -i "s#path: /mnt/models#path: $MOUNT_DIR#g" {} \;
# find . -name '*.yaml' -type f -exec sed -i "s#path: /mnt/models#path: $MOUNT_DIR#g" {} \;
find . -name '*.yaml' -type f -exec sed -i "s#path: /mnt/opea-models#path: $MOUNT_DIR#g" {} \;
# replace the repository "image: opea/*" with "image: ${IMAGE_REPO}opea/"
find . -name '*.yaml' -type f -exec sed -i "s#image: opea/*#image: ${IMAGE_REPO}opea/#g" {} \;
find . -name '*.yaml' -type f -exec sed -i "s#image: \"opea/*#image: \"${IMAGE_REPO}opea/#g" {} \;
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/scripts/e2e/gmc_xeon_test.sh
Original file line number Diff line number Diff line change
Expand Up @@ -83,7 +83,7 @@ function validate_chatqna() {
echo "Checking response results, make sure the output is reasonable. "
local status=false
if [[ -f $LOG_PATH/curl_chatqna.log ]] && \
[[ $(grep -c "billion" $LOG_PATH/curl_chatqna.log) != 0 ]]; then
[[ $(grep -c "[DONE]" $LOG_PATH/curl_chatqna.log) != 0 ]]; then
status=true
fi
if [ $status == false ]; then
Expand Down
5 changes: 5 additions & 0 deletions microservices-connector/config/samples/chatQnA_gaudi.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -23,6 +23,7 @@ spec:
serviceName: embedding-svc
config:
endpoint: /v1/embeddings
TEI_EMBEDDING_ENDPOINT: tei-embedding-gaudi-svc
- name: TeiEmbeddingGaudi
internalService:
serviceName: tei-embedding-gaudi-svc
Expand All @@ -33,6 +34,8 @@ spec:
serviceName: retriever-svc
config:
endpoint: /v1/retrieval
REDIS_URL: redis-vector-db
TEI_EMBEDDING_ENDPOINT: tei-embedding-gaudi-svc
- name: VectorDB
internalService:
serviceName: redis-vector-db
Expand All @@ -43,6 +46,7 @@ spec:
serviceName: reranking-svc
config:
endpoint: /v1/reranking
TEI_RERANKING_ENDPOINT: tei-reranking-svc
- name: TeiReranking
internalService:
serviceName: tei-reranking-svc
Expand All @@ -55,6 +59,7 @@ spec:
serviceName: llm-svc
config:
endpoint: /v1/chat/completions
TGI_LLM_ENDPOINT: tgi-gaudi-svc
- name: TgiGaudi
internalService:
serviceName: tgi-gaudi-svc
Expand Down
5 changes: 5 additions & 0 deletions microservices-connector/config/samples/chatQnA_xeon.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -23,6 +23,7 @@ spec:
serviceName: embedding-svc
config:
endpoint: /v1/embeddings
TEI_EMBEDDING_ENDPOINT: tei-embedding-svc
- name: TeiEmbedding
internalService:
serviceName: tei-embedding-svc
Expand All @@ -33,6 +34,8 @@ spec:
serviceName: retriever-svc
config:
endpoint: /v1/retrieval
REDIS_URL: redis-vector-db
TEI_EMBEDDING_ENDPOINT: tei-embedding-svc
- name: VectorDB
internalService:
serviceName: redis-vector-db
Expand All @@ -43,6 +46,7 @@ spec:
serviceName: reranking-svc
config:
endpoint: /v1/reranking
TEI_RERANKING_ENDPOINT: tei-reranking-svc
- name: TeiReranking
internalService:
serviceName: tei-reranking-svc
Expand All @@ -55,6 +59,7 @@ spec:
serviceName: llm-svc
config:
endpoint: /v1/chat/completions
TGI_LLM_ENDPOINT: tgi-service-m
- name: Tgi
internalService:
serviceName: tgi-service-m
Expand Down
3 changes: 2 additions & 1 deletion microservices-connector/config/samples/codegen_gaudi.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -24,10 +24,11 @@ spec:
serviceName: llm-service
config:
endpoint: /v1/chat/completions
TGI_LLM_ENDPOINT: tgi-gaudi-svc
- name: TgiGaudi
internalService:
serviceName: tgi-gaudi-svc
config:
LLM_MODEL_ID: ise-uiuc/Magicoder-S-DS-6.7B
MODEL_ID: ise-uiuc/Magicoder-S-DS-6.7B
endpoint: /generate
isDownstreamService: true
3 changes: 2 additions & 1 deletion microservices-connector/config/samples/codegen_xeon.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -24,10 +24,11 @@ spec:
serviceName: llm-service
config:
endpoint: /v1/chat/completions
TGI_LLM_ENDPOINT: tgi-service
- name: Tgi
internalService:
serviceName: tgi-service
config:
LLM_MODEL_ID: ise-uiuc/Magicoder-S-DS-6.7B
MODEL_ID: ise-uiuc/Magicoder-S-DS-6.7B
endpoint: /generate
isDownstreamService: true
3 changes: 2 additions & 1 deletion microservices-connector/config/samples/codetrans_gaudi.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -24,10 +24,11 @@ spec:
serviceName: codetrans-service
config:
endpoint: /v1/chat/completions
TGI_LLM_ENDPOINT: tgi-gaudi-svc
- name: TgiGaudi
internalService:
serviceName: tgi-gaudi-svc
config:
LLM_MODEL_ID: HuggingFaceH4/mistral-7b-grok
MODEL_ID: HuggingFaceH4/mistral-7b-grok
endpoint: /generate
isDownstreamService: true
3 changes: 2 additions & 1 deletion microservices-connector/config/samples/codetrans_xeon.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -24,10 +24,11 @@ spec:
serviceName: codetrans-service
config:
endpoint: /v1/chat/completions
TGI_LLM_ENDPOINT: tgi-service
- name: Tgi
internalService:
serviceName: tgi-service
config:
LLM_MODEL_ID: HuggingFaceH4/mistral-7b-grok
MODEL_ID: HuggingFaceH4/mistral-7b-grok
endpoint: /generate
isDownstreamService: true
1 change: 1 addition & 0 deletions microservices-connector/config/samples/docsum_gaudi.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -27,6 +27,7 @@ spec:
HUGGING_FACE_HUB_TOKEN: <HUGGING_FACE_HUB_TOKEN>
HF_TOKEN: <HF_TOKEN>
PORT: "9009"
TGI_LLM_ENDPOINT: tgi-gaudi-svc
- name: TgiGaudi
internalService:
serviceName: tgi-gaudi-svc
Expand Down
1 change: 1 addition & 0 deletions microservices-connector/config/samples/docsum_xeon.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -27,6 +27,7 @@ spec:
HUGGING_FACE_HUB_TOKEN: <HUGGING_FACE_HUB_TOKEN>
HF_TOKEN: <HF_TOKEN>
PORT: "9009"
TGI_LLM_ENDPOINT: tgi-svc
- name: Tgi
internalService:
serviceName: tgi-svc
Expand Down
Loading
Loading