refine readme for reorg (#782)

Signed-off-by: Xinyao Wang <xinyao.wang@intel.com>
Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
This commit is contained in:
XinyaoWa
2024-09-11 14:57:29 +08:00
committed by GitHub
parent d97882ec8e
commit d2bab99835
28 changed files with 60 additions and 60 deletions

View File

@@ -7,14 +7,14 @@
## Deploy On Xeon
```
cd GenAIExamples/AudioQnA/kubernetes/manifests/xeon
cd GenAIExamples/AudioQnA/kubernetes/intel/cpu/xeon/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" audioqna.yaml
kubectl apply -f audioqna.yaml
```
## Deploy On Gaudi
```
cd GenAIExamples/AudioQnA/kubernetes/manifests/gaudi
cd GenAIExamples/AudioQnA/kubernetes/intel/hpu/gaudi/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" audioqna.yaml
kubectl apply -f audioqna.yaml

View File

@@ -123,7 +123,7 @@ Currently we support two ways of deploying ChatQnA services with docker compose:
docker pull opea/chatqna-conversation-ui:latest
```
2. Using the docker images `built from source`: [Guide](docker/xeon/README.md)
2. Using the docker images `built from source`: [Guide](docker_compose/intel/cpu/xeon/README.md)
> Note: The **opea/chatqna-without-rerank:latest** docker image has not been published yet, users need to build this docker image from source.
@@ -139,7 +139,7 @@ By default, the embedding, reranking and LLM models are set to a default value a
Change the `xxx_MODEL_ID` in `docker/xxx/set_env.sh` for your needs.
For customers with proxy issues, the models from [ModelScope](https://www.modelscope.cn/models) are also supported in ChatQnA. Refer to [this readme](docker/xeon/README.md) for details.
For customers with proxy issues, the models from [ModelScope](https://www.modelscope.cn/models) are also supported in ChatQnA. Refer to [this readme](docker_compose/intel/cpu/xeon/README.md) for details.
### Setup Environment Variable
@@ -202,11 +202,11 @@ Refer to the [Xeon Guide](./docker_compose/intel/cpu/xeon/README.md) for more in
### Deploy ChatQnA on NVIDIA GPU
```bash
cd GenAIExamples/ChatQnA/docker/gpu/
cd GenAIExamples/ChatQnA/docker_compose/nvidia/gpu/
docker compose up -d
```
Refer to the [NVIDIA GPU Guide](./docker/gpu/README.md) for more instructions on building docker images from source.
Refer to the [NVIDIA GPU Guide](./docker_compose/nvidia/gpu/README.md) for more instructions on building docker images from source.
### Deploy ChatQnA into Kubernetes on Xeon & Gaudi with GMC
@@ -214,7 +214,7 @@ Refer to the [Kubernetes Guide](./kubernetes/intel/README_gmc.md) for instructio
### Deploy ChatQnA into Kubernetes on Xeon & Gaudi without GMC
Refer to the [Kubernetes Guide](./kubernetes/kubernetes/intel/README.md) for instructions on deploying ChatQnA into Kubernetes on Xeon & Gaudi without GMC.
Refer to the [Kubernetes Guide](./kubernetes/intel/README.md) for instructions on deploying ChatQnA into Kubernetes on Xeon & Gaudi without GMC.
### Deploy ChatQnA into Kubernetes using Helm Chart
@@ -224,7 +224,7 @@ Refer to the [ChatQnA helm chart](https://github.com/opea-project/GenAIInfra/tre
### Deploy ChatQnA on AI PC
Refer to the [AI PC Guide](./docker/aipc/README.md) for instructions on deploying ChatQnA on AI PC.
Refer to the [AI PC Guide](./docker_compose/intel/cpu/aipc/README.md) for instructions on deploying ChatQnA on AI PC.
### Deploy ChatQnA on Red Hat OpenShift Container Platform (RHOCP)

View File

@@ -159,7 +159,7 @@ Note: Please replace with `host_ip` with you external IP address, do not use loc
> Before running the docker compose command, you need to be in the folder that has the docker compose yaml file
```bash
cd GenAIExamples/ChatQnA/docker/aipc/
cd GenAIExamples/ChatQnA/docker_compose/intel/cpu/aipc/
docker compose up -d
# let ollama service runs

View File

@@ -147,7 +147,7 @@ cd ..
Build frontend Docker image via below command:
```bash
cd GenAIExamples/ChatQnA/
cd GenAIExamples/ChatQnA/ui
docker build --no-cache -t opea/chatqna-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile .
cd ../../../..
```

View File

@@ -85,7 +85,7 @@ docker build --no-cache -t opea/retriever-qdrant:latest --build-arg https_proxy=
### 3. Build Rerank Image
```bash
docker build --no-cache -t opea/reranking-tei:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/reranks/tei/Dockerfile .
docker build --no-cache -t opea/reranking-tei:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/reranks/tei/Dockerfile .`
```
### 4. Build LLM Image
@@ -117,7 +117,7 @@ cd ../../..
Build frontend Docker image via below command:
```bash
cd GenAIExamples/ChatQnA/
cd GenAIExamples/ChatQnA/ui
docker build --no-cache -t opea/chatqna-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile .
cd ../../../..
```

View File

@@ -128,7 +128,7 @@ cd ../..
Construct the frontend Docker image using the command below:
```bash
cd GenAIExamples/ChatQnA/
cd GenAIExamples/ChatQnA/ui
docker build --no-cache -t opea/chatqna-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile .
cd ../../../..
```
@@ -150,7 +150,7 @@ cd ../../../..
To fortify AI initiatives in production, Guardrails microservice can secure model inputs and outputs, building Trustworthy, Safe, and Secure LLM-based Applications.
```bash
cd GenAIExamples/ChatQnA/docker
cd GenAIComps
docker build -t opea/guardrails-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/guardrails/llama_guard/langchain/Dockerfile .
cd ../../..
```

View File

@@ -59,7 +59,7 @@ cd ../../..
Construct the frontend Docker image using the command below:
```bash
cd GenAIExamples/ChatQnA/
cd GenAIExamples/ChatQnA/ui
docker build --no-cache -t opea/chatqna-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile .
cd ../../../..
```
@@ -132,7 +132,7 @@ Note: Please replace with `host_ip` with you external IP address, do **NOT** use
### Start all the services Docker Containers
```bash
cd GenAIExamples/ChatQnA/docker/gpu/
cd GenAIExamples/ChatQnA/docker_compose/nvidia/gpu/
docker compose up -d
```

View File

@@ -11,7 +11,7 @@
## Deploy On Xeon
```
cd GenAIExamples/ChatQnA/kubernetes/manifests/xeon
cd GenAIExamples/ChatQnA/kubernetes/intel/cpu/xeon/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" chatqna.yaml
kubectl apply -f chatqna.yaml
@@ -20,7 +20,7 @@ kubectl apply -f chatqna.yaml
## Deploy On Gaudi
```
cd GenAIExamples/ChatQnA/kubernetes/manifests/gaudi
cd GenAIExamples/ChatQnA/kubernetes/intel/hpu/gaudi/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" chatqna.yaml
kubectl apply -f chatqna.yaml

View File

@@ -166,7 +166,7 @@ case "$1" in
if [ $ret -ne 0 ]; then
exit $ret
fi
pushd ChatQnA/kubernetes/manifests/gaudi
pushd ChatQnA/kubernetes/intel/hpu/gaudi/manifests
set +e
install_and_validate_chatqna_guardrail
popd

View File

@@ -166,7 +166,7 @@ case "$1" in
if [ $ret -ne 0 ]; then
exit $ret
fi
pushd ChatQnA/kubernetes/manifests/xeon
pushd ChatQnA/kubernetes/intel/cpu/xeon/manifests
set +e
install_and_validate_chatqna_guardrail
popd

View File

@@ -12,7 +12,7 @@
## Deploy On Xeon
```
cd GenAIExamples/CodeGen/kubernetes/manifests/xeon
cd GenAIExamples/CodeGen/kubernetes/intel/cpu/xeon/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
export MODEL_ID="meta-llama/CodeLlama-7b-hf"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" codegen.yaml
@@ -23,7 +23,7 @@ kubectl apply -f codegen.yaml
## Deploy On Gaudi
```
cd GenAIExamples/CodeGen/kubernetes/manifests/gaudi
cd GenAIExamples/CodeGen/kubernetes/intel/hpu/gaudi/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" codegen.yaml
kubectl apply -f codegen.yaml

View File

@@ -17,7 +17,7 @@ Before deploying the react-codegen.yaml file, ensure that you have the following
```
# You may set the HUGGINGFACEHUB_API_TOKEN via method:
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
cd GenAIExamples/CodeGen/kubernetes/manifests/xeon/ui/
cd GenAIExamples/CodeGen/kubernetes/intel/cpu/xeon/manifests/ui/
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" react-codegen.yaml
```
b. Set the proxies based on your network configuration

View File

@@ -21,7 +21,7 @@ Change the `MODEL_ID` in `codetrans.yaml` for your needs.
## Deploy On Xeon
```bash
cd GenAIExamples/CodeTrans/kubernetes/manifests/xeon
cd GenAIExamples/CodeTrans/kubernetes/intel/cpu/xeon/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" codetrans.yaml
kubectl apply -f codetrans.yaml
@@ -30,7 +30,7 @@ kubectl apply -f codetrans.yaml
## Deploy On Gaudi
```bash
cd GenAIExamples/CodeTrans/kubernetes/manifests/gaudi
cd GenAIExamples/CodeTrans/kubernetes/intel/hpu/gaudi/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" codetrans.yaml
kubectl apply -f codetrans.yaml

View File

@@ -21,7 +21,7 @@ Currently we support two ways of deploying Document Summarization services with
docker pull opea/docsum:latest
```
2. Start services using the docker images `built from source`: [Guide](./docker)
2. Start services using the docker images `built from source`: [Guide](./docker_compose)
### Required Models

View File

@@ -11,7 +11,7 @@
## Deploy On Xeon
```
cd GenAIExamples/DocSum/kubernetes/manifests/xeon
cd GenAIExamples/DocSum/kubernetes/intel/cpu/xeon/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" docsum.yaml
kubectl apply -f docsum.yaml
@@ -20,7 +20,7 @@ kubectl apply -f docsum.yaml
## Deploy On Gaudi
```
cd GenAIExamples/DocSum/kubernetes/manifests/gaudi
cd GenAIExamples/DocSum/kubernetes/intel/hpu/gaudi/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" docsum.yaml
kubectl apply -f docsum.yaml

View File

@@ -16,7 +16,7 @@ Before deploying the react-docsum.yaml file, ensure that you have the following
```
# You may set the HUGGINGFACEHUB_API_TOKEN via method:
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
cd GenAIExamples/DocSum/kubernetes/manifests/xeon/ui/
cd GenAIExamples/DocSum/kubernetes/intel/cpu/xeon/manifests/ui/
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" react-docsum.yaml
```
b. Set the proxies based on your network configuration

View File

@@ -17,7 +17,7 @@ If use gated models, you also need to provide [huggingface token](https://huggin
## Deploy On Xeon
```
cd GenAIExamples/FaqGen/kubernetes/manifests/xeon
cd GenAIExamples/FaqGen/kubernetes/intel/cpu/xeon/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" faqgen.yaml
kubectl apply -f faqgen.yaml
@@ -26,7 +26,7 @@ kubectl apply -f faqgen.yaml
## Deploy On Gaudi
```
cd GenAIExamples/FaqGen/kubernetes/manifests/gaudi
cd GenAIExamples/FaqGen/kubernetes/intel/hpu/gaudi/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" faqgen.yaml
kubectl apply -f faqgen.yaml

View File

@@ -16,7 +16,7 @@ Before deploying the react-faqgen.yaml file, ensure that you have the following
```
# You may set the HUGGINGFACEHUB_API_TOKEN via method:
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
cd GenAIExamples/FaqGen/kubernetes/manifests/xeon/ui/
cd GenAIExamples/FaqGen/kubernetes/intel/cpu/xeon/manifests/ui/
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" react-faqgen.yaml
```
b. Set the proxies based on your network configuration

View File

@@ -20,4 +20,4 @@ Refer to the [Keycloak Configuration Guide](./docker_compose/intel/cpu/xeon/keyc
Refer to the [Xeon Guide](./docker_compose/intel/cpu/xeon/README.md) for more instructions on building docker images from source and running the application via docker compose.
Refer to the [Xeon Kubernetes Guide](./kubernetes/manifests/README.md) for more instruction on deploying the application via kubernetes.
Refer to the [Xeon Kubernetes Guide](./kubernetes/intel/README.md) for more instruction on deploying the application via kubernetes.

View File

@@ -27,7 +27,7 @@ To begin with, ensure that you have following prerequisites in place:
```
# You may set the HUGGINGFACEHUB_API_TOKEN via method:
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
cd GenAIExamples/ProductivitySuite/kubernetes/manifests/xeon/
cd GenAIExamples/ProductivitySuite/kubernetes/intel/cpu/xeon/manifests/
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" *.yaml
```
@@ -48,7 +48,7 @@ To begin with, ensure that you have following prerequisites in place:
## Deploying ProductivitySuite
You can use yaml files in xeon folder to deploy ProductivitySuite with reactUI.
```
cd GenAIExamples/ProductivitySuite/kubernetes/manifests/xeon/
cd GenAIExamples/ProductivitySuite/kubernetes/intel/cpu/xeon/manifests/
kubectl apply -f *.yaml
```

View File

@@ -37,17 +37,17 @@ Deployment are based on released docker images by default, check [docker image l
#### Deploy Examples
| Use Case | Docker Compose<br/>Deployment on Xeon | Docker Compose<br/>Deployment on Gaudi | Kubernetes with GMC | Kubernetes with Manifests | Kubernetes with Helm Charts |
| ----------- | ------------------------------------------------------------------------ | -------------------------------------------------------------------------- | ------------------------------------------------------------------ | ---------------------------------------------------------------- | ------------------------------------------------------------------------------------------------------------------ |
| ChatQnA | [Xeon Instructions](ChatQnA/docker_compose/intel/cpu/xeon/README.md) | [Gaudi Instructions](ChatQnA/docker_compose/intel/hpu/gaudi/README.md) | [ChatQnA with GMC](ChatQnA/kubernetes/intel/README_gmc.md) | [ChatQnA with Manifests](ChatQnA/kubernetes/intel/README.md) | [ChatQnA with Helm Charts](https://github.com/opea-project/GenAIInfra/tree/main/helm-charts/chatqna/README.md) |
| CodeGen | [Xeon Instructions](CodeGen/docker_compose/intel/cpu/xeon/README.md) | [Gaudi Instructions](CodeGen/docker_compose/intel/hpu/gaudi/README.md) | [CodeGen with GMC](CodeGen/kubernetes/intel/README_gmc.md) | [CodeGen with Manifests](CodeGen/kubernetes/intel/README.md) | [CodeGen with Helm Charts](https://github.com/opea-project/GenAIInfra/tree/main/helm-charts/codegen/README.md) |
| CodeTrans | [Xeon Instructions](CodeTrans/docker_compose/intel/cpu/xeon/README.md) | [Gaudi Instructions](CodeTrans/docker_compose/intel/hpu/gaudi/README.md) | [CodeTrans with GMC](CodeTrans/kubernetes/intel/README_gmc.md) | [CodeTrans with Manifests](CodeTrans/kubernetes/intel/README.md) | [CodeTrans with Helm Charts](https://github.com/opea-project/GenAIInfra/tree/main/helm-charts/codetrans/README.md) |
| DocSum | [Xeon Instructions](DocSum/docker_compose/intel/cpu/xeon/README.md) | [Gaudi Instructions](DocSum/docker_compose/intel/hpu/gaudi/README.md) | [DocSum with GMC](DocSum/kubernetes/intel/README_gmc.md) | [DocSum with Manifests](DocSum/kubernetes/intel/README.md) | [DocSum with Helm Charts](https://github.com/opea-project/GenAIInfra/tree/main/helm-charts/docsum/README.md) |
| SearchQnA | [Xeon Instructions](SearchQnA/docker_compose/intel/cpu/xeon/README.md) | [Gaudi Instructions](SearchQnA/docker_compose/intel/hpu/gaudi/README.md) | [SearchQnA with GMC](SearchQnA/kubernetes/intel/README_gmc.md) | Not Supported | Not Supported |
| FaqGen | [Xeon Instructions](FaqGen/docker_compose/intel/cpu/xeon/README.md) | [Gaudi Instructions](FaqGen/docker_compose/intel/hpu/gaudi/README.md) | [FaqGen with GMC](FaqGen/kubernetes/intel/README_gmc.md) | [FaqGen with Manifests](FaqGen/kubernetes/intel/README.md) | Not Supported |
| Translation | [Xeon Instructions](Translation/docker_compose/intel/cpu/xeon/README.md) | [Gaudi Instructions](Translation/docker_compose/intel/hpu/gaudi/README.md) | [Translation with GMC](Translation/kubernetes/intel/README_gmc.md) | Not Supported | Not Supported |
| AudioQnA | [Xeon Instructions](AudioQnA/docker_compose/intel/cpu/xeon/README.md) | [Gaudi Instructions](AudioQnA/docker_compose/intel/hpu/gaudi/README.md) | [AudioQnA with GMC](AudioQnA/kubernetes/intel/README_gmc.md) | [AudioQnA with Manifests](AudioQnA/kubernetes/intel/README.md) | Not Supported |
| VisualQnA | [Xeon Instructions](VisualQnA/docker_compose/intel/cpu/xeon/README.md) | [Gaudi Instructions](VisualQnA/docker_compose/intel/hpu/gaudi/README.md) | [VisualQnA with GMC](VisualQnA/kubernetes/intel/README_gmc.md) | [VisualQnA with Manifests](VisualQnA/kubernetes/intel/README.md) | Not Supported |
| Use Case | Docker Compose<br/>Deployment on Xeon | Docker Compose<br/>Deployment on Gaudi | Kubernetes with Manifests | Kubernetes with Helm Charts | Kubernetes with GMC |
| ----------- | ------------------------------------------------------------------------ | -------------------------------------------------------------------------- | ---------------------------------------------------------------- | ------------------------------------------------------------------------------------------------------------------ | ------------------------------------------------------------------ |
| ChatQnA | [Xeon Instructions](ChatQnA/docker_compose/intel/cpu/xeon/README.md) | [Gaudi Instructions](ChatQnA/docker_compose/intel/hpu/gaudi/README.md) | [ChatQnA with Manifests](ChatQnA/kubernetes/intel/README.md) | [ChatQnA with Helm Charts](https://github.com/opea-project/GenAIInfra/tree/main/helm-charts/chatqna/README.md) | [ChatQnA with GMC](ChatQnA/kubernetes/intel/README_gmc.md) |
| CodeGen | [Xeon Instructions](CodeGen/docker_compose/intel/cpu/xeon/README.md) | [Gaudi Instructions](CodeGen/docker_compose/intel/hpu/gaudi/README.md) | [CodeGen with Manifests](CodeGen/kubernetes/intel/README.md) | [CodeGen with Helm Charts](https://github.com/opea-project/GenAIInfra/tree/main/helm-charts/codegen/README.md) | [CodeGen with GMC](CodeGen/kubernetes/intel/README_gmc.md) |
| CodeTrans | [Xeon Instructions](CodeTrans/docker_compose/intel/cpu/xeon/README.md) | [Gaudi Instructions](CodeTrans/docker_compose/intel/hpu/gaudi/README.md) | [CodeTrans with Manifests](CodeTrans/kubernetes/intel/README.md) | [CodeTrans with Helm Charts](https://github.com/opea-project/GenAIInfra/tree/main/helm-charts/codetrans/README.md) | [CodeTrans with GMC](CodeTrans/kubernetes/intel/README_gmc.md) |
| DocSum | [Xeon Instructions](DocSum/docker_compose/intel/cpu/xeon/README.md) | [Gaudi Instructions](DocSum/docker_compose/intel/hpu/gaudi/README.md) | [DocSum with Manifests](DocSum/kubernetes/intel/README.md) | [DocSum with Helm Charts](https://github.com/opea-project/GenAIInfra/tree/main/helm-charts/docsum/README.md) | [DocSum with GMC](DocSum/kubernetes/intel/README_gmc.md) |
| SearchQnA | [Xeon Instructions](SearchQnA/docker_compose/intel/cpu/xeon/README.md) | [Gaudi Instructions](SearchQnA/docker_compose/intel/hpu/gaudi/README.md) | Not Supported | Not Supported | [SearchQnA with GMC](SearchQnA/kubernetes/intel/README_gmc.md) |
| FaqGen | [Xeon Instructions](FaqGen/docker_compose/intel/cpu/xeon/README.md) | [Gaudi Instructions](FaqGen/docker_compose/intel/hpu/gaudi/README.md) | [FaqGen with Manifests](FaqGen/kubernetes/intel/README.md) | Not Supported | [FaqGen with GMC](FaqGen/kubernetes/intel/README_gmc.md) |
| Translation | [Xeon Instructions](Translation/docker_compose/intel/cpu/xeon/README.md) | [Gaudi Instructions](Translation/docker_compose/intel/hpu/gaudi/README.md) | Not Supported | Not Supported | [Translation with GMC](Translation/kubernetes/intel/README_gmc.md) |
| AudioQnA | [Xeon Instructions](AudioQnA/docker_compose/intel/cpu/xeon/README.md) | [Gaudi Instructions](AudioQnA/docker_compose/intel/hpu/gaudi/README.md) | [AudioQnA with Manifests](AudioQnA/kubernetes/intel/README.md) | Not Supported | [AudioQnA with GMC](AudioQnA/kubernetes/intel/README_gmc.md) |
| VisualQnA | [Xeon Instructions](VisualQnA/docker_compose/intel/cpu/xeon/README.md) | [Gaudi Instructions](VisualQnA/docker_compose/intel/hpu/gaudi/README.md) | [VisualQnA with Manifests](VisualQnA/kubernetes/intel/README.md) | Not Supported | [VisualQnA with GMC](VisualQnA/kubernetes/intel/README_gmc.md) |
## Supported Examples

View File

@@ -6,7 +6,7 @@ Rerank model finetuning is the process of further training rerank model on a dat
### Deploy Rerank Model Finetuning Service on Xeon
Refer to the [Xeon Guide](./docker/xeon/README.md) for detail.
Refer to the [Xeon Guide](./docker_compose/intel/cpu/xeon/README.md) for detail.
## Consume Rerank Model Finetuning Service

View File

@@ -32,7 +32,7 @@ Currently we support two ways of deploying SearchQnA services with docker compos
docker pull opea/searchqna:latest
```
2. Start services using the docker images `built from source`: [Guide](./docker)
2. Start services using the docker images `built from source`: [Guide](./docker_compose)
### Setup Environment Variable

View File

@@ -1,10 +0,0 @@
[
{
"src": "/extreme_ironing.jpg",
"prompt": "what is unusual about this image?"
},
{
"src": "/waterview.jpg",
"prompt": "what are the things I should be cautious about when I visit here?"
}
]

View File

@@ -8,14 +8,14 @@
## Deploy On Xeon
```
cd GenAIExamples/visualqna/kubernetes/manifests/xeon
cd GenAIExamples/visualqna/kubernetes/intel/cpu/xeon/manifests
kubectl apply -f visualqna.yaml
```
## Deploy On Gaudi
```
cd GenAIExamples/visualqna/kubernetes/manifests/gaudi
cd GenAIExamples/visualqna/kubernetes/intel/hpu/gaudi/manifests
kubectl apply -f visualqna.yaml
```

View File

Before

Width:  |  Height:  |  Size: 61 KiB

After

Width:  |  Height:  |  Size: 61 KiB

View File

@@ -0,0 +1,10 @@
[
{
"src": "/extreme_ironing.jpg",
"prompt": "what is unusual about this image?"
},
{
"src": "/waterview.jpg",
"prompt": "what are the things I should be cautious about when I visit here?"
}
]

View File

Before

Width:  |  Height:  |  Size: 93 KiB

After

Width:  |  Height:  |  Size: 93 KiB