Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Make all xeon tgi image version consistent #851

Merged
merged 14 commits into from
Sep 24, 2024

Conversation

zhlsunshine
Copy link

Description

Make all xeon tgi image version consistent

Issues

n/a.

Type of change

List the type of change like below. Please delete options that are not relevant.

  • Bug fix (non-breaking change which fixes an issue)
  • New feature (non-breaking change which adds new functionality)
  • Breaking change (fix or feature that would break existing design and interface)
  • Others (enhancement, documentation, validation, etc.)

Dependencies

n/a.

Tests

n/a.

…-cpu as the tgi image on xeon.

Signed-off-by: zhlsunshine <[email protected]>
@zhlsunshine
Copy link
Author

Hi @lvliang-intel, as we discussed, let's use ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu as the tgi image on xeon for all examples.

@yinghu5 yinghu5 requested review from yinghu5 and chensuyue and removed request for yinghu5 September 19, 2024 14:12
@chensuyue chensuyue modified the milestone: v1.1 Sep 19, 2024
@zhlsunshine
Copy link
Author

Hi @lvliang-intel, I replace the ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu with ghcr.io/huggingface/text-generation-inference:latest-intel-cpu

Copy link
Collaborator

@lianhao lianhao left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

why the CI manifest test is missing? We need them to make sure it's working

@zhlsunshine
Copy link
Author

Hi @lvliang-intel @lianhao @chensuyue, after discussed with Intel Xeon TGI image support engineer, we'd better to use the specified version of TGI image, such as ghcr.io/huggingface/text-generation-inference:sha-e4201f4-intel-cpu, instead ghcr.io/huggingface/text-generation-inference:latest-intel-cpu which would update every day.

@chensuyue
Copy link
Collaborator

Please resolve the conflicts.

Signed-off-by: zhlsunshine <[email protected]>
@zhlsunshine
Copy link
Author

Please resolve the conflicts.

Hi @chensuyue sure, done!

@chensuyue chensuyue merged commit 954a220 into opea-project:main Sep 24, 2024
12 checks passed
wangkl2 pushed a commit to wangkl2/GenAIExamples that referenced this pull request Dec 11, 2024
* draft a demo code for memory.

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

* add agent short-term memory with langgraph checkpoint.

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

* add save long-term memory func.

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

* add save long-term memory func.

* add timeout for llm response.

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

* fix ut with adding -e HABANA_VISIBLE_DEVICES=all.

---------

Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants