From 264759d85ab5af7f8ff13a5185d0debede78d031 Mon Sep 17 00:00:00 2001 From: XinyaoWa Date: Thu, 12 Sep 2024 17:52:06 +0800 Subject: [PATCH] fix path bug for reorg (#801) Signed-off-by: Xinyao Wang --- AgentQnA/tests/4_launch_and_validate_agent_openai.sh | 2 +- AudioQnA/tests/test_compose_on_gaudi.sh | 4 ++-- AudioQnA/tests/test_compose_on_xeon.sh | 4 ++-- ChatQnA/README.md | 2 +- ChatQnA/docker_compose/intel/cpu/xeon/README_qdrant.md | 2 +- CodeGen/README.md | 2 +- CodeGen/{ => docker_compose}/set_env.sh | 0 CodeTrans/README.md | 4 ++-- CodeTrans/docker_compose/intel/cpu/xeon/README.md | 3 ++- CodeTrans/docker_compose/intel/hpu/gaudi/README.md | 3 ++- CodeTrans/{ => docker_compose}/set_env.sh | 0 DocSum/README.md | 4 ++-- DocSum/{ => docker_compose}/set_env.sh | 0 SearchQnA/README.md | 2 +- SearchQnA/{ => docker_compose}/set_env.sh | 0 15 files changed, 17 insertions(+), 15 deletions(-) rename CodeGen/{ => docker_compose}/set_env.sh (100%) rename CodeTrans/{ => docker_compose}/set_env.sh (100%) rename DocSum/{ => docker_compose}/set_env.sh (100%) rename SearchQnA/{ => docker_compose}/set_env.sh (100%) diff --git a/AgentQnA/tests/4_launch_and_validate_agent_openai.sh b/AgentQnA/tests/4_launch_and_validate_agent_openai.sh index fb9a8949f..b3220c09d 100644 --- a/AgentQnA/tests/4_launch_and_validate_agent_openai.sh +++ b/AgentQnA/tests/4_launch_and_validate_agent_openai.sh @@ -16,7 +16,7 @@ function start_agent_and_api_server() { docker run -d --runtime=runc --name=kdd-cup-24-crag-service -p=8080:8000 docker.io/aicrowd/kdd-cup-24-crag-mock-api:v0 echo "Starting Agent services" - cd $WORKDIR/GenAIExamples/AgentQnA/docker/openai + cd $WORKDIR/GenAIExamples/AgentQnA/docker_compose/intel/cpu/xeon bash launch_agent_service_openai.sh } diff --git a/AudioQnA/tests/test_compose_on_gaudi.sh b/AudioQnA/tests/test_compose_on_gaudi.sh index c8fde79b2..221888f91 100644 --- a/AudioQnA/tests/test_compose_on_gaudi.sh +++ b/AudioQnA/tests/test_compose_on_gaudi.sh @@ -45,7 +45,7 @@ function start_services() { export TTS_SERVICE_PORT=3002 export LLM_SERVICE_PORT=3007 - # sed -i "s/backend_address/$ip_address/g" $WORKPATH/docker/ui/svelte/.env + # sed -i "s/backend_address/$ip_address/g" $WORKPATH/ui/svelte/.env # Start Docker Containers docker compose up -d > ${LOG_PATH}/start_services_with_compose.log @@ -91,7 +91,7 @@ function validate_megaservice() { } #function validate_frontend() { -# cd $WORKPATH/docker/ui/svelte +# cd $WORKPATH/ui/svelte # local conda_env_name="OPEA_e2e" # export PATH=${HOME}/miniforge3/bin/:$PATH ## conda remove -n ${conda_env_name} --all -y diff --git a/AudioQnA/tests/test_compose_on_xeon.sh b/AudioQnA/tests/test_compose_on_xeon.sh index b37440b1b..998d5f362 100644 --- a/AudioQnA/tests/test_compose_on_xeon.sh +++ b/AudioQnA/tests/test_compose_on_xeon.sh @@ -44,7 +44,7 @@ function start_services() { export TTS_SERVICE_PORT=3002 export LLM_SERVICE_PORT=3007 - # sed -i "s/backend_address/$ip_address/g" $WORKPATH/docker/ui/svelte/.env + # sed -i "s/backend_address/$ip_address/g" $WORKPATH/ui/svelte/.env # Start Docker Containers docker compose up -d > ${LOG_PATH}/start_services_with_compose.log @@ -81,7 +81,7 @@ function validate_megaservice() { } #function validate_frontend() { -# cd $WORKPATH/docker/ui/svelte +# cd $WORKPATH/ui/svelte # local conda_env_name="OPEA_e2e" # export PATH=${HOME}/miniforge3/bin/:$PATH ## conda remove -n ${conda_env_name} --all -y diff --git a/ChatQnA/README.md b/ChatQnA/README.md index c0e809f7f..dd0fb159b 100644 --- a/ChatQnA/README.md +++ b/ChatQnA/README.md @@ -152,7 +152,7 @@ By default, the embedding, reranking and LLM models are set to a default value a | Reranking | BAAI/bge-reranker-base | | LLM | Intel/neural-chat-7b-v3-3 | -Change the `xxx_MODEL_ID` in `docker/xxx/set_env.sh` for your needs. +Change the `xxx_MODEL_ID` in `docker_compose/xxx/set_env.sh` for your needs. For customers with proxy issues, the models from [ModelScope](https://www.modelscope.cn/models) are also supported in ChatQnA. Refer to [this readme](docker_compose/intel/cpu/xeon/README.md) for details. diff --git a/ChatQnA/docker_compose/intel/cpu/xeon/README_qdrant.md b/ChatQnA/docker_compose/intel/cpu/xeon/README_qdrant.md index 9eb23ecc5..25ba15c3f 100644 --- a/ChatQnA/docker_compose/intel/cpu/xeon/README_qdrant.md +++ b/ChatQnA/docker_compose/intel/cpu/xeon/README_qdrant.md @@ -107,7 +107,7 @@ To construct the Mega Service, we utilize the [GenAIComps](https://github.com/op ```bash git clone https://github.com/opea-project/GenAIExamples.git -cd GenAIExamples/ChatQnA/docker +cd GenAIExamples/ChatQnA/ docker build --no-cache -t opea/chatqna:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . cd ../../.. ``` diff --git a/CodeGen/README.md b/CodeGen/README.md index 259008731..bc93ff473 100644 --- a/CodeGen/README.md +++ b/CodeGen/README.md @@ -67,7 +67,7 @@ To set up environment variables for deploying ChatQnA services, follow these ste 3. Set up other environment variables: ```bash - source ./docker/set_env.sh + source ./docker_compose/set_env.sh ``` ### Deploy CodeGen using Docker diff --git a/CodeGen/set_env.sh b/CodeGen/docker_compose/set_env.sh similarity index 100% rename from CodeGen/set_env.sh rename to CodeGen/docker_compose/set_env.sh diff --git a/CodeTrans/README.md b/CodeTrans/README.md index b5bf56207..f40911971 100644 --- a/CodeTrans/README.md +++ b/CodeTrans/README.md @@ -30,7 +30,7 @@ By default, the LLM model is set to a default value as listed below: | ------- | ----------------------------- | | LLM | HuggingFaceH4/mistral-7b-grok | -Change the `LLM_MODEL_ID` in `docker/set_env.sh` for your needs. +Change the `LLM_MODEL_ID` in `docker_compose/set_env.sh` for your needs. ### Setup Environment Variable @@ -58,7 +58,7 @@ To set up environment variables for deploying Code Translation services, follow 3. Set up other environment variables: ```bash - source ./docker/set_env.sh + source ./docker_compose/set_env.sh ``` ### Deploy with Docker diff --git a/CodeTrans/docker_compose/intel/cpu/xeon/README.md b/CodeTrans/docker_compose/intel/cpu/xeon/README.md index 6293566db..81618bf5c 100755 --- a/CodeTrans/docker_compose/intel/cpu/xeon/README.md +++ b/CodeTrans/docker_compose/intel/cpu/xeon/README.md @@ -92,7 +92,8 @@ Change the `LLM_MODEL_ID` below for your needs. 3. Set up other environment variables: ```bash - source ../set_env.sh + cd GenAIExamples/CodeTrans/docker_compose + source ./set_env.sh ``` ### Start Microservice Docker Containers diff --git a/CodeTrans/docker_compose/intel/hpu/gaudi/README.md b/CodeTrans/docker_compose/intel/hpu/gaudi/README.md index 11fbd22ad..0ce5f8879 100755 --- a/CodeTrans/docker_compose/intel/hpu/gaudi/README.md +++ b/CodeTrans/docker_compose/intel/hpu/gaudi/README.md @@ -84,7 +84,8 @@ Change the `LLM_MODEL_ID` below for your needs. 3. Set up other environment variables: ```bash - source ../set_env.sh + cd GenAIExamples/CodeTrans/docker_compose + source ./set_env.sh ``` ### Start Microservice Docker Containers diff --git a/CodeTrans/set_env.sh b/CodeTrans/docker_compose/set_env.sh similarity index 100% rename from CodeTrans/set_env.sh rename to CodeTrans/docker_compose/set_env.sh diff --git a/DocSum/README.md b/DocSum/README.md index 540bb2558..6714b2b8d 100644 --- a/DocSum/README.md +++ b/DocSum/README.md @@ -25,7 +25,7 @@ Currently we support two ways of deploying Document Summarization services with ### Required Models -We set default model as "Intel/neural-chat-7b-v3-3", change "LLM_MODEL_ID" in "set_env.sh" if you want to use other models. +We set default model as "Intel/neural-chat-7b-v3-3", change "LLM_MODEL_ID" in "docker_compose/set_env.sh" if you want to use other models. ``` export LLM_MODEL_ID="Intel/neural-chat-7b-v3-3" @@ -57,7 +57,7 @@ To set up environment variables for deploying Document Summarization services, f 3. Set up other environment variables: ```bash - source ./docker/set_env.sh + source ./docker_compose/set_env.sh ``` ### Deploy using Docker diff --git a/DocSum/set_env.sh b/DocSum/docker_compose/set_env.sh similarity index 100% rename from DocSum/set_env.sh rename to DocSum/docker_compose/set_env.sh diff --git a/SearchQnA/README.md b/SearchQnA/README.md index 433c46996..5e9656b80 100644 --- a/SearchQnA/README.md +++ b/SearchQnA/README.md @@ -60,7 +60,7 @@ To set up environment variables for deploying SearchQnA services, follow these s 3. Set up other environment variables: ```bash - source ./docker/set_env.sh + source ./docker_compose/set_env.sh ``` ### Deploy SearchQnA on Gaudi diff --git a/SearchQnA/set_env.sh b/SearchQnA/docker_compose/set_env.sh similarity index 100% rename from SearchQnA/set_env.sh rename to SearchQnA/docker_compose/set_env.sh