-
Notifications
You must be signed in to change notification settings - Fork 213
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug] can NOT launch ChatQnA on xeon using K8S manifest files #752
Comments
lianhao
added a commit
to lianhao/GenAIExamples
that referenced
this issue
Sep 6, 2024
- Update ChatQnA/CodeGen/CodeTrans/DocSum k8s manifest to avoid requiring creating directory for cache model. - Add chatqna-guardrails manifest files. - Fix bug opea-project#752 introduced by PR opea-project#669 Signed-off-by: Lianhao Lu <[email protected]>
2 tasks
yongfengdu
pushed a commit
that referenced
this issue
Sep 6, 2024
- Update ChatQnA/CodeGen/CodeTrans/DocSum k8s manifest to avoid requiring creating directory for cache model. - Add chatqna-guardrails manifest files. - Fix bug #752 introduced by PR #669 Signed-off-by: Lianhao Lu <[email protected]>
xuechendi
pushed a commit
to xuechendi/GenAIExamples
that referenced
this issue
Sep 9, 2024
- Update ChatQnA/CodeGen/CodeTrans/DocSum k8s manifest to avoid requiring creating directory for cache model. - Add chatqna-guardrails manifest files. - Fix bug opea-project#752 introduced by PR opea-project#669 Signed-off-by: Lianhao Lu <[email protected]>
dmsuehir
pushed a commit
to dmsuehir/GenAIExamples
that referenced
this issue
Sep 11, 2024
- Update ChatQnA/CodeGen/CodeTrans/DocSum k8s manifest to avoid requiring creating directory for cache model. - Add chatqna-guardrails manifest files. - Fix bug opea-project#752 introduced by PR opea-project#669 Signed-off-by: Lianhao Lu <[email protected]>
wangkl2
pushed a commit
to wangkl2/GenAIExamples
that referenced
this issue
Dec 11, 2024
* Update Agent vllm client codes and test Signed-off-by: Chendi.Xue <[email protected]> * Update README with vllm related update Signed-off-by: Chendi.Xue <[email protected]> * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * remove unnecessary env Signed-off-by: Chendi.Xue <[email protected]> * support plan_execute with vllm Signed-off-by: Chendi.Xue <[email protected]> * Update README and test llama vllm support Signed-off-by: Chendi.Xue <[email protected]> * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * update vllm_port to 8086 avoid conflict Signed-off-by: Chendi.Xue <[email protected]> --------- Signed-off-by: Chendi.Xue <[email protected]> Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Priority
P1-Stopper
OS type
Ubuntu
Hardware type
Xeon-SPR
Installation method
Deploy method
Running nodes
Single Node
What's the version?
The version with PR #669 merged.
Description
Launch ChatQnA on xeon by the following command:
Then check the pod status by the followings:
nginx pod keeps crashing. The root cause of this that in nginx configuration, the upstream URL is hardcoded, which doesn't support install ChatQnA in non-default k8s namespace.
Reproduce steps
kubectl create ns lianhao-test
kubectl install -n lianhao-test -f ChatQnA/kubernetes/manifests/xeon/chatqna.yaml
kubectl -n lianhao-test get pod
Raw log
No response
The text was updated successfully, but these errors were encountered: