Explain Default Model in ChatQnA and CodeTrans READMEs (#694)
* explain default model in CodeTrans READMEs Signed-off-by: letonghan <letong.han@intel.com> * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * explain default model in ChatQnA READMEs Signed-off-by: letonghan <letong.han@intel.com> * add required models Signed-off-by: letonghan <letong.han@intel.com> * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci --------- Signed-off-by: letonghan <letong.han@intel.com> Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
This commit is contained in:
@@ -22,6 +22,16 @@ Currently we support two ways of deploying Code Translation services on docker:
|
||||
|
||||
2. Start services using the docker images `built from source`: [Guide](./docker)
|
||||
|
||||
### Required Models
|
||||
|
||||
By default, the LLM model is set to a default value as listed below:
|
||||
|
||||
| Service | Model |
|
||||
| ------- | ----------------------------- |
|
||||
| LLM | HuggingFaceH4/mistral-7b-grok |
|
||||
|
||||
Change the `LLM_MODEL_ID` in `docker/set_env.sh` for your needs.
|
||||
|
||||
### Setup Environment Variable
|
||||
|
||||
To set up environment variables for deploying Code Translation services, follow these steps:
|
||||
|
||||
@@ -42,9 +42,17 @@ Then run the command `docker images`, you will have the following Docker Images:
|
||||
|
||||
## 🚀 Start Microservices
|
||||
|
||||
### Setup Environment Variables
|
||||
### Required Models
|
||||
|
||||
Since the `compose.yaml` will consume some environment variables, you need to setup them in advance as below. Notice that the `LLM_MODEL_ID` indicates the LLM model used for TGI service.
|
||||
By default, the LLM model is set to a default value as listed below:
|
||||
|
||||
| Service | Model |
|
||||
| ------- | ----------------------------- |
|
||||
| LLM | HuggingFaceH4/mistral-7b-grok |
|
||||
|
||||
Change the `LLM_MODEL_ID` below for your needs.
|
||||
|
||||
### Setup Environment Variables
|
||||
|
||||
```bash
|
||||
export no_proxy=${your_no_proxy}
|
||||
|
||||
@@ -50,9 +50,17 @@ Then run the command `docker images`, you will have the following Docker Images:
|
||||
|
||||
## 🚀 Start Microservices
|
||||
|
||||
### Setup Environment Variables
|
||||
### Required Models
|
||||
|
||||
Since the `compose.yaml` will consume some environment variables, you need to setup them in advance as below. Notice that the `LLM_MODEL_ID` indicates the LLM model used for TGI service.
|
||||
By default, the LLM model is set to a default value as listed below:
|
||||
|
||||
| Service | Model |
|
||||
| ------- | ----------------------------- |
|
||||
| LLM | HuggingFaceH4/mistral-7b-grok |
|
||||
|
||||
Change the `LLM_MODEL_ID` below for your needs.
|
||||
|
||||
### Setup Environment Variables
|
||||
|
||||
```bash
|
||||
export no_proxy=${your_no_proxy}
|
||||
|
||||
@@ -7,9 +7,20 @@ Please install GMC in your Kubernetes cluster, if you have not already done so,
|
||||
If you have only Intel Xeon machines you could use the codetrans_xeon.yaml file or if you have a Gaudi cluster you could use codetrans_gaudi.yaml
|
||||
In the below example we illustrate on Xeon.
|
||||
|
||||
## Required Models
|
||||
|
||||
By default, the LLM model is set to a default value as listed below:
|
||||
|
||||
|Service |Model |
|
||||
|---------|-------------------------|
|
||||
|LLM |HuggingFaceH4/mistral-7b-grok|
|
||||
|
||||
Change the `MODEL_ID` in `codetrans_xeon.yaml` for your needs.
|
||||
|
||||
## Deploy the RAG application
|
||||
|
||||
1. Create the desired namespace if it does not already exist and deploy the application
|
||||
|
||||
```bash
|
||||
export APP_NAMESPACE=CT
|
||||
kubectl create ns $APP_NAMESPACE
|
||||
|
||||
@@ -8,6 +8,16 @@
|
||||
|
||||
> You need to make sure you have created the directory `/mnt/opea-models` to save the cached model on the node where the CodeTrans workload is running. Otherwise, you need to modify the `codetrans.yaml` file to change the `model-volume` to a directory that exists on the node.
|
||||
|
||||
## Required Models
|
||||
|
||||
By default, the LLM model is set to a default value as listed below:
|
||||
|
||||
|Service |Model |
|
||||
|---------|-------------------------|
|
||||
|LLM |HuggingFaceH4/mistral-7b-grok|
|
||||
|
||||
Change the `MODEL_ID` in `codetrans.yaml` for your needs.
|
||||
|
||||
## Deploy On Xeon
|
||||
|
||||
```bash
|
||||
|
||||
Reference in New Issue
Block a user