Skip to content

Commit

Permalink
Explain Default Model in ChatQnA and CodeTrans READMEs (#694)
Browse files Browse the repository at this point in the history
* explain default model in CodeTrans READMEs

Signed-off-by: letonghan <[email protected]>

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

* explain default model in ChatQnA READMEs

Signed-off-by: letonghan <[email protected]>

* add required models

Signed-off-by: letonghan <[email protected]>

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

---------

Signed-off-by: letonghan <[email protected]>
Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
  • Loading branch information
letonghan and pre-commit-ci[bot] authored Aug 29, 2024
1 parent 6a679ba commit 2a2ff45
Show file tree
Hide file tree
Showing 10 changed files with 112 additions and 5 deletions.
12 changes: 12 additions & 0 deletions ChatQnA/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -121,6 +121,18 @@ Currently we support two ways of deploying ChatQnA services with docker compose:

2. Start services using the docker images `built from source`: [Guide](./docker)

### Required Models

By default, the embedding, reranking and LLM models are set to a default value as listed below:

| Service | Model |
| --------- | ------------------------- |
| Embedding | BAAI/bge-base-en-v1.5 |
| Reranking | BAAI/bge-reranker-base |
| LLM | Intel/neural-chat-7b-v3-3 |

Change the `xxx_MODEL_ID` in `docker/xxx/set_env.sh` for your needs.

### Setup Environment Variable

To set up environment variables for deploying ChatQnA services, follow these steps:
Expand Down
12 changes: 12 additions & 0 deletions ChatQnA/docker/gaudi/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -159,6 +159,18 @@ If Guardrails docker image is built, you will find one more image:

## 🚀 Start MicroServices and MegaService

### Required Models

By default, the embedding, reranking and LLM models are set to a default value as listed below:

| Service | Model |
| --------- | ------------------------- |
| Embedding | BAAI/bge-base-en-v1.5 |
| Reranking | BAAI/bge-reranker-base |
| LLM | Intel/neural-chat-7b-v3-3 |

Change the `xxx_MODEL_ID` below for your needs.

### Setup Environment Variables

Since the `compose.yaml` will consume some environment variables, you need to setup them in advance as below.
Expand Down
12 changes: 12 additions & 0 deletions ChatQnA/docker/gpu/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -87,6 +87,18 @@ Then run the command `docker images`, you will have the following 7 Docker Image

## 🚀 Start MicroServices and MegaService

### Required Models

By default, the embedding, reranking and LLM models are set to a default value as listed below:

| Service | Model |
| --------- | ------------------------- |
| Embedding | BAAI/bge-base-en-v1.5 |
| Reranking | BAAI/bge-reranker-base |
| LLM | Intel/neural-chat-7b-v3-3 |

Change the `xxx_MODEL_ID` below for your needs.

### Setup Environment Variables

Since the `compose.yaml` will consume some environment variables, you need to setup them in advance as below.
Expand Down
14 changes: 13 additions & 1 deletion ChatQnA/docker/xeon/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -161,6 +161,18 @@ Then run the command `docker images`, you will have the following 7 Docker Image

## 🚀 Start Microservices

### Required Models

By default, the embedding, reranking and LLM models are set to a default value as listed below:

| Service | Model |
| --------- | ------------------------- |
| Embedding | BAAI/bge-base-en-v1.5 |
| Reranking | BAAI/bge-reranker-base |
| LLM | Intel/neural-chat-7b-v3-3 |

Change the `xxx_MODEL_ID` below for your needs.

### Setup Environment Variables

Since the `compose.yaml` will consume some environment variables, you need to setup them in advance as below.
Expand All @@ -183,7 +195,7 @@ export your_hf_api_token="Your_Huggingface_API_Token"

**Append the value of the public IP address to the no_proxy list**

```
```bash
export your_no_proxy=${your_no_proxy},"External_Public_IP"
```

Expand Down
12 changes: 12 additions & 0 deletions ChatQnA/docker/xeon/README_qdrant.md
Original file line number Diff line number Diff line change
Expand Up @@ -148,6 +148,18 @@ Then run the command `docker images`, you will have the following 7 Docker Image

## 🚀 Start Microservices

### Required Models

By default, the embedding, reranking and LLM models are set to a default value as listed below:

| Service | Model |
| --------- | ------------------------- |
| Embedding | BAAI/bge-base-en-v1.5 |
| Reranking | BAAI/bge-reranker-base |
| LLM | Intel/neural-chat-7b-v3-3 |

Change the `xxx_MODEL_ID` below for your needs.

### Setup Environment Variables

Since the `compose.yaml` will consume some environment variables, you need to setup them in advance as below.
Expand Down
10 changes: 10 additions & 0 deletions CodeTrans/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,16 @@ Currently we support two ways of deploying Code Translation services on docker:

2. Start services using the docker images `built from source`: [Guide](./docker)

### Required Models

By default, the LLM model is set to a default value as listed below:

| Service | Model |
| ------- | ----------------------------- |
| LLM | HuggingFaceH4/mistral-7b-grok |

Change the `LLM_MODEL_ID` in `docker/set_env.sh` for your needs.

### Setup Environment Variable

To set up environment variables for deploying Code Translation services, follow these steps:
Expand Down
12 changes: 10 additions & 2 deletions CodeTrans/docker/gaudi/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -42,9 +42,17 @@ Then run the command `docker images`, you will have the following Docker Images:

## 🚀 Start Microservices

### Setup Environment Variables
### Required Models

By default, the LLM model is set to a default value as listed below:

| Service | Model |
| ------- | ----------------------------- |
| LLM | HuggingFaceH4/mistral-7b-grok |

Since the `compose.yaml` will consume some environment variables, you need to setup them in advance as below. Notice that the `LLM_MODEL_ID` indicates the LLM model used for TGI service.
Change the `LLM_MODEL_ID` below for your needs.

### Setup Environment Variables

```bash
export no_proxy=${your_no_proxy}
Expand Down
12 changes: 10 additions & 2 deletions CodeTrans/docker/xeon/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -50,9 +50,17 @@ Then run the command `docker images`, you will have the following Docker Images:

## 🚀 Start Microservices

### Setup Environment Variables
### Required Models

By default, the LLM model is set to a default value as listed below:

| Service | Model |
| ------- | ----------------------------- |
| LLM | HuggingFaceH4/mistral-7b-grok |

Since the `compose.yaml` will consume some environment variables, you need to setup them in advance as below. Notice that the `LLM_MODEL_ID` indicates the LLM model used for TGI service.
Change the `LLM_MODEL_ID` below for your needs.

### Setup Environment Variables

```bash
export no_proxy=${your_no_proxy}
Expand Down
11 changes: 11 additions & 0 deletions CodeTrans/kubernetes/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,9 +7,20 @@ Please install GMC in your Kubernetes cluster, if you have not already done so,
If you have only Intel Xeon machines you could use the codetrans_xeon.yaml file or if you have a Gaudi cluster you could use codetrans_gaudi.yaml
In the below example we illustrate on Xeon.

## Required Models

By default, the LLM model is set to a default value as listed below:

|Service |Model |
|---------|-------------------------|
|LLM |HuggingFaceH4/mistral-7b-grok|

Change the `MODEL_ID` in `codetrans_xeon.yaml` for your needs.

## Deploy the RAG application

1. Create the desired namespace if it does not already exist and deploy the application

```bash
export APP_NAMESPACE=CT
kubectl create ns $APP_NAMESPACE
Expand Down
10 changes: 10 additions & 0 deletions CodeTrans/kubernetes/manifests/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,16 @@
> You need to make sure you have created the directory `/mnt/opea-models` to save the cached model on the node where the CodeTrans workload is running. Otherwise, you need to modify the `codetrans.yaml` file to change the `model-volume` to a directory that exists on the node.
## Required Models

By default, the LLM model is set to a default value as listed below:

|Service |Model |
|---------|-------------------------|
|LLM |HuggingFaceH4/mistral-7b-grok|

Change the `MODEL_ID` in `codetrans.yaml` for your needs.

## Deploy On Xeon

```bash
Expand Down

0 comments on commit 2a2ff45

Please sign in to comment.