Skip to content

Commit d487093

Browse files
authored
Add default model in readme for FaqGen and DocSum (#693)
* update default model in readme for DocSum Signed-off-by: Xinyao Wang <[email protected]>
1 parent e6f5d13 commit d487093

File tree

7 files changed

+47
-4
lines changed

7 files changed

+47
-4
lines changed

DocSum/README.md

Lines changed: 10 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,16 @@ Currently we support two ways of deploying Document Summarization services with
2323

2424
2. Start services using the docker images `built from source`: [Guide](./docker)
2525

26+
### Required Models
27+
28+
We set default model as "Intel/neural-chat-7b-v3-3", change "LLM_MODEL_ID" in "set_env.sh" if you want to use other models.
29+
30+
```
31+
export LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
32+
```
33+
34+
If use gated models, you also need to provide [huggingface token](https://huggingface.co/docs/hub/security-tokens) to "HUGGINGFACEHUB_API_TOKEN" environment variable.
35+
2636
### Setup Environment Variable
2737

2838
To set up environment variables for deploying Document Summarization services, follow these steps:

DocSum/docker/gaudi/README.md

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -64,6 +64,11 @@ Then run the command `docker images`, you will have the following Docker Images:
6464

6565
## 🚀 Start Microservices and MegaService
6666

67+
### Required Models
68+
69+
We set default model as "Intel/neural-chat-7b-v3-3", change "LLM_MODEL_ID" in following setting if you want to use other models.
70+
If use gated models, you also need to provide [huggingface token](https://huggingface.co/docs/hub/security-tokens) to "HUGGINGFACEHUB_API_TOKEN" environment variable.
71+
6772
### Setup Environment Variables
6873

6974
Since the `compose.yaml` will consume some environment variables, you need to setup them in advance as below.

DocSum/docker/xeon/README.md

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -73,6 +73,11 @@ Then run the command `docker images`, you will have the following Docker Images:
7373

7474
## 🚀 Start Microservices and MegaService
7575

76+
### Required Models
77+
78+
We set default model as "Intel/neural-chat-7b-v3-3", change "LLM_MODEL_ID" in following Environment Variables setting if you want to use other models.
79+
If use gated models, you also need to provide [huggingface token](https://huggingface.co/docs/hub/security-tokens) to "HUGGINGFACEHUB_API_TOKEN" environment variable.
80+
7681
### Setup Environment Variables
7782

7883
Since the `compose.yaml` will consume some environment variables, you need to setup them in advance as below.

DocSum/kubernetes/README.md

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,9 @@ These will be available on Docker Hub soon, simplifying installation.
2020
This involves deploying the application pipeline custom resource. You can use docsum_xeon.yaml if you have just a Xeon cluster or docsum_gaudi.yaml if you have a Gaudi cluster.
2121

2222
1. Setup Environment variables. These are specific to the user. Skip the proxy settings if you are not operating behind one.
23-
23+
24+
We use "Intel/neural-chat-7b-v3-3" as an example. If you want to use other models, change "LLM_MODEL_ID" in following setting and change "MODEL_ID" in manifests yaml file.
25+
2426
```bash
2527
export no_proxy=${your_no_proxy}
2628
export http_proxy=${your_http_proxy}

FaqGen/docker/gaudi/README.md

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -64,6 +64,12 @@ Then run the command `docker images`, you will have the following Docker Images:
6464

6565
## 🚀 Start Microservices and MegaService
6666

67+
### Required Models
68+
69+
We set default model as "meta-llama/Meta-Llama-3-8B-Instruct", change "LLM_MODEL_ID" in following Environment Variables setting if you want to use other models.
70+
71+
If use gated models, you also need to provide [huggingface token](https://huggingface.co/docs/hub/security-tokens) to "HUGGINGFACEHUB_API_TOKEN" environment variable.
72+
6773
### Setup Environment Variables
6874

6975
Since the `compose.yaml` will consume some environment variables, you need to setup them in advance as below.
@@ -72,7 +78,7 @@ Since the `compose.yaml` will consume some environment variables, you need to se
7278
export no_proxy=${your_no_proxy}
7379
export http_proxy=${your_http_proxy}
7480
export https_proxy=${your_http_proxy}
75-
export LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
81+
export LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
7682
export TGI_LLM_ENDPOINT="http://${your_ip}:8008"
7783
export HUGGINGFACEHUB_API_TOKEN=${your_hf_api_token}
7884
export MEGA_SERVICE_HOST_IP=${host_ip}

FaqGen/docker/xeon/README.md

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -63,6 +63,12 @@ Then run the command `docker images`, you will have the following Docker Images:
6363

6464
## 🚀 Start Microservices and MegaService
6565

66+
### Required Models
67+
68+
We set default model as "meta-llama/Meta-Llama-3-8B-Instruct", change "LLM_MODEL_ID" in following Environment Variables setting if you want to use other models.
69+
70+
If use gated models, you also need to provide [huggingface token](https://huggingface.co/docs/hub/security-tokens) to "HUGGINGFACEHUB_API_TOKEN" environment variable.
71+
6672
### Setup Environment Variables
6773

6874
Since the `compose.yaml` will consume some environment variables, you need to setup them in advance as below.
@@ -71,7 +77,7 @@ Since the `compose.yaml` will consume some environment variables, you need to se
7177
export no_proxy=${your_no_proxy}
7278
export http_proxy=${your_http_proxy}
7379
export https_proxy=${your_http_proxy}
74-
export LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
80+
export LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
7581
export TGI_LLM_ENDPOINT="http://${your_ip}:8008"
7682
export HUGGINGFACEHUB_API_TOKEN=${your_hf_api_token}
7783
export MEGA_SERVICE_HOST_IP=${host_ip}

FaqGen/kubernetes/manifests/README.md

Lines changed: 10 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,16 @@
33
> [NOTE]
44
> The following values must be set before you can deploy:
55
> HUGGINGFACEHUB_API_TOKEN
6-
> You can also customize the "MODEL_ID" and "model-volume"
6+
> You can also customize the "MODEL_ID" and "model-volume".
7+
8+
## Required Models
9+
We set "meta-llama/Meta-Llama-3-8B-Instruct" as default model, if you want to use other models, change arguments "--model-id" in `xeon/faqgen.yaml` or `gaudi/faqgen.yaml`.
10+
```
11+
- --model-id
12+
- 'meta-llama/Meta-Llama-3-8B-Instruct'
13+
```
14+
15+
If use gated models, you also need to provide [huggingface token](https://huggingface.co/docs/hub/security-tokens) to "HUGGINGFACEHUB_API_TOKEN" environment variable.
716

817
## Deploy On Xeon
918

0 commit comments

Comments
 (0)