From 5ac77f78dabe773870a68e9d77c3b0554ab8035c Mon Sep 17 00:00:00 2001 From: "chen, suyue" Date: Thu, 8 Aug 2024 13:39:38 +0800 Subject: [PATCH] fix faqgen on xeon test scripts (#552) Signed-off-by: chensuyue --- .github/workflows/VisualQnA.yml | 50 ----------------------------- FaqGen/docker/xeon/compose.yaml | 2 +- FaqGen/tests/test_faqgen_on_xeon.sh | 5 +-- 3 files changed, 4 insertions(+), 53 deletions(-) delete mode 100644 .github/workflows/VisualQnA.yml diff --git a/.github/workflows/VisualQnA.yml b/.github/workflows/VisualQnA.yml deleted file mode 100644 index da0194829..000000000 --- a/.github/workflows/VisualQnA.yml +++ /dev/null @@ -1,50 +0,0 @@ -# Copyright (C) 2024 Intel Corporation -# SPDX-License-Identifier: Apache-2.0 - -name: VisualQnA-test - -on: - pull_request_target: - branches: [main] - types: [opened, reopened, ready_for_review, synchronize] # added `ready_for_review` since draft is skipped - paths: - - VisualQnA/** - - "!**.md" - - "!**/ui/**" - - .github/workflows/VisualQnA.yml - workflow_dispatch: - -# If there is a new commit, the previous jobs will be canceled -concurrency: - group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }} - cancel-in-progress: true - -jobs: - VisualQnA: - runs-on: aise-cluster - strategy: - matrix: - job_name: ["basic"] - fail-fast: false - steps: - - name: Clean Up Working Directory - run: sudo rm -rf ${{github.workspace}}/* - - - name: Checkout out Repo - uses: actions/checkout@v4 - with: - ref: "refs/pull/${{ github.event.number }}/merge" - - - name: Run Test - env: - HUGGINGFACEHUB_API_TOKEN: ${{ secrets.HUGGINGFACEHUB_API_TOKEN }} - run: | - cd ${{ github.workspace }}/VisualQnA/tests - bash test_${{ matrix.job_name }}_inference.sh - - - name: Publish pipeline artifact - if: ${{ !cancelled() }} - uses: actions/upload-artifact@v4 - with: - name: ${{ matrix.job_name }} - path: ${{ github.workspace }}/VisualQnA/tests/*.log diff --git a/FaqGen/docker/xeon/compose.yaml b/FaqGen/docker/xeon/compose.yaml index b8f775962..00f2d0ed0 100644 --- a/FaqGen/docker/xeon/compose.yaml +++ b/FaqGen/docker/xeon/compose.yaml @@ -6,7 +6,7 @@ version: "3.8" services: tgi-service: image: ghcr.io/huggingface/text-generation-inference:1.4 - container_name: tgi_xeon_server + container_name: tgi-xeon-server ports: - "8008:80" environment: diff --git a/FaqGen/tests/test_faqgen_on_xeon.sh b/FaqGen/tests/test_faqgen_on_xeon.sh index f72baf503..b0b694f21 100755 --- a/FaqGen/tests/test_faqgen_on_xeon.sh +++ b/FaqGen/tests/test_faqgen_on_xeon.sh @@ -9,9 +9,10 @@ LOG_PATH="$WORKPATH/tests" ip_address=$(hostname -I | awk '{print $1}') function build_docker_images() { - cd $WORKPATH/../../ + cd $WORKPATH git clone https://github.com/opea-project/GenAIComps.git + cd GenAIComps docker build --no-cache -t opea/llm-faqgen-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/faq-generation/tgi/Dockerfile . @@ -94,7 +95,7 @@ function validate_microservices() { "${ip_address}:8008/generate" \ "generated_text" \ "tgi-service" \ - "tgi_xeon_server" \ + "tgi-xeon-server" \ '{"inputs":"What is Deep Learning?","parameters":{"max_new_tokens":17, "do_sample": true}}' # llm microservice