Compare commits

...

1 Commits

Author SHA1 Message Date
ZePan110
3e29a76fc4 update secrets token name
Signed-off-by: ZePan110 <ze.pan@intel.com>
2025-05-21 17:12:00 +08:00
137 changed files with 267 additions and 267 deletions

View File

@@ -55,7 +55,7 @@ jobs:
- name: Run tests - name: Run tests
id: run-test id: run-test
env: env:
HUGGINGFACEHUB_API_TOKEN: ${{ secrets.HUGGINGFACEHUB_API_TOKEN }} HUGGINGFACEHUB_API_TOKEN: ${{ secrets.HF_TOKEN }}
GOOGLE_CSE_ID: ${{ secrets.GOOGLE_CSE_ID }} GOOGLE_CSE_ID: ${{ secrets.GOOGLE_CSE_ID }}
GOOGLE_API_KEY: ${{ secrets.GOOGLE_API_KEY }} GOOGLE_API_KEY: ${{ secrets.GOOGLE_API_KEY }}
run: | run: |

View File

@@ -165,8 +165,8 @@ jobs:
env: env:
GOOGLE_CSE_ID: ${{ secrets.GOOGLE_CSE_ID }} GOOGLE_CSE_ID: ${{ secrets.GOOGLE_CSE_ID }}
GOOGLE_API_KEY: ${{ secrets.GOOGLE_API_KEY }} GOOGLE_API_KEY: ${{ secrets.GOOGLE_API_KEY }}
HUGGINGFACEHUB_API_TOKEN: ${{ secrets.HUGGINGFACEHUB_API_TOKEN }} HUGGINGFACEHUB_API_TOKEN: ${{ secrets.HF_TOKEN }}
HFTOKEN: ${{ secrets.HUGGINGFACEHUB_API_TOKEN }} HFTOKEN: ${{ secrets.HF_TOKEN }}
value_file: ${{ matrix.value_file }} value_file: ${{ matrix.value_file }}
run: | run: |
set -xe set -xe

View File

@@ -160,8 +160,8 @@ jobs:
- name: Run test - name: Run test
shell: bash shell: bash
env: env:
HUGGINGFACEHUB_API_TOKEN: ${{ secrets.HUGGINGFACEHUB_API_TOKEN }} HUGGINGFACEHUB_API_TOKEN: ${{ secrets.HF_TOKEN }}
HF_TOKEN: ${{ secrets.HUGGINGFACEHUB_API_TOKEN }} HF_TOKEN: ${{ secrets.HF_TOKEN }}
GOOGLE_CSE_ID: ${{ secrets.GOOGLE_CSE_ID }} GOOGLE_CSE_ID: ${{ secrets.GOOGLE_CSE_ID }}
GOOGLE_API_KEY: ${{ secrets.GOOGLE_API_KEY }} GOOGLE_API_KEY: ${{ secrets.GOOGLE_API_KEY }}
PINECONE_KEY: ${{ secrets.PINECONE_KEY }} PINECONE_KEY: ${{ secrets.PINECONE_KEY }}

View File

@@ -13,8 +13,8 @@ services:
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
TGI_LLM_ENDPOINT: "http://${ip_address}:${TGI_SERVICE_PORT}" TGI_LLM_ENDPOINT: "http://${ip_address}:${TGI_SERVICE_PORT}"
HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
shm_size: 32g shm_size: 32g
devices: devices:
- /dev/kfd:/dev/kfd - /dev/kfd:/dev/kfd
@@ -42,7 +42,7 @@ services:
with_memory: false with_memory: false
recursion_limit: ${recursion_limit_worker} recursion_limit: ${recursion_limit_worker}
llm_engine: tgi llm_engine: tgi
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
llm_endpoint_url: ${LLM_ENDPOINT_URL} llm_endpoint_url: ${LLM_ENDPOINT_URL}
model: ${LLM_MODEL_ID} model: ${LLM_MODEL_ID}
temperature: ${temperature} temperature: ${temperature}
@@ -76,7 +76,7 @@ services:
use_hints: false use_hints: false
recursion_limit: ${recursion_limit_worker} recursion_limit: ${recursion_limit_worker}
llm_engine: vllm llm_engine: vllm
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
llm_endpoint_url: ${LLM_ENDPOINT_URL} llm_endpoint_url: ${LLM_ENDPOINT_URL}
model: ${LLM_MODEL_ID} model: ${LLM_MODEL_ID}
temperature: ${temperature} temperature: ${temperature}
@@ -104,7 +104,7 @@ services:
with_memory: true with_memory: true
recursion_limit: ${recursion_limit_supervisor} recursion_limit: ${recursion_limit_supervisor}
llm_engine: tgi llm_engine: tgi
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
llm_endpoint_url: ${LLM_ENDPOINT_URL} llm_endpoint_url: ${LLM_ENDPOINT_URL}
model: ${LLM_MODEL_ID} model: ${LLM_MODEL_ID}
temperature: ${temperature} temperature: ${temperature}

View File

@@ -10,8 +10,8 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
WILM_USE_TRITON_FLASH_ATTENTION: 0 WILM_USE_TRITON_FLASH_ATTENTION: 0
@@ -46,7 +46,7 @@ services:
with_memory: false with_memory: false
recursion_limit: ${recursion_limit_worker} recursion_limit: ${recursion_limit_worker}
llm_engine: vllm llm_engine: vllm
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
llm_endpoint_url: ${LLM_ENDPOINT_URL} llm_endpoint_url: ${LLM_ENDPOINT_URL}
model: ${LLM_MODEL_ID} model: ${LLM_MODEL_ID}
temperature: ${temperature} temperature: ${temperature}
@@ -80,7 +80,7 @@ services:
use_hints: false use_hints: false
recursion_limit: ${recursion_limit_worker} recursion_limit: ${recursion_limit_worker}
llm_engine: vllm llm_engine: vllm
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
llm_endpoint_url: ${LLM_ENDPOINT_URL} llm_endpoint_url: ${LLM_ENDPOINT_URL}
model: ${LLM_MODEL_ID} model: ${LLM_MODEL_ID}
temperature: ${temperature} temperature: ${temperature}
@@ -108,7 +108,7 @@ services:
with_memory: true with_memory: true
recursion_limit: ${recursion_limit_supervisor} recursion_limit: ${recursion_limit_supervisor}
llm_engine: vllm llm_engine: vllm
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
llm_endpoint_url: ${LLM_ENDPOINT_URL} llm_endpoint_url: ${LLM_ENDPOINT_URL}
model: ${LLM_MODEL_ID} model: ${LLM_MODEL_ID}
temperature: ${temperature} temperature: ${temperature}

View File

@@ -19,7 +19,7 @@ export CRAG_SERVER_PORT="18114"
export WORKPATH=$(dirname "$PWD") export WORKPATH=$(dirname "$PWD")
export WORKDIR=${WORKPATH}/../../../ export WORKDIR=${WORKPATH}/../../../
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export LLM_MODEL_ID="Intel/neural-chat-7b-v3-3" export LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
export HF_CACHE_DIR="./data" export HF_CACHE_DIR="./data"
export MODEL_CACHE="./data" export MODEL_CACHE="./data"
@@ -39,7 +39,7 @@ export CRAG_SERVER=http://${ip_address}:${CRAG_SERVER_PORT}
export WORKER_AGENT_URL="http://${ip_address}:${WORKER_RAG_AGENT_PORT}/v1/chat/completions" export WORKER_AGENT_URL="http://${ip_address}:${WORKER_RAG_AGENT_PORT}/v1/chat/completions"
export SQL_AGENT_URL="http://${ip_address}:${WORKER_SQL_AGENT_PORT}/v1/chat/completions" export SQL_AGENT_URL="http://${ip_address}:${WORKER_SQL_AGENT_PORT}/v1/chat/completions"
export HF_CACHE_DIR=${HF_CACHE_DIR} export HF_CACHE_DIR=${HF_CACHE_DIR}
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export no_proxy=${no_proxy} export no_proxy=${no_proxy}
export http_proxy=${http_proxy} export http_proxy=${http_proxy}
export https_proxy=${https_proxy} export https_proxy=${https_proxy}

View File

@@ -19,7 +19,7 @@ export CRAG_SERVER_PORT="18114"
export WORKPATH=$(dirname "$PWD") export WORKPATH=$(dirname "$PWD")
export WORKDIR=${WORKPATH}/../../../ export WORKDIR=${WORKPATH}/../../../
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export VLLM_LLM_MODEL_ID="Intel/neural-chat-7b-v3-3" export VLLM_LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
export HF_CACHE_DIR="./data" export HF_CACHE_DIR="./data"
export MODEL_CACHE="./data" export MODEL_CACHE="./data"
@@ -40,7 +40,7 @@ export CRAG_SERVER=http://${ip_address}:${CRAG_SERVER_PORT}
export WORKER_AGENT_URL="http://${ip_address}:${WORKER_RAG_AGENT_PORT}/v1/chat/completions" export WORKER_AGENT_URL="http://${ip_address}:${WORKER_RAG_AGENT_PORT}/v1/chat/completions"
export SQL_AGENT_URL="http://${ip_address}:${WORKER_SQL_AGENT_PORT}/v1/chat/completions" export SQL_AGENT_URL="http://${ip_address}:${WORKER_SQL_AGENT_PORT}/v1/chat/completions"
export HF_CACHE_DIR=${HF_CACHE_DIR} export HF_CACHE_DIR=${HF_CACHE_DIR}
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export no_proxy=${no_proxy} export no_proxy=${no_proxy}
export http_proxy=${http_proxy} export http_proxy=${http_proxy}
export https_proxy=${https_proxy} export https_proxy=${https_proxy}

View File

@@ -20,8 +20,8 @@ export CRAG_SERVER_PORT="18114"
export WORKPATH=$(dirname "$PWD") export WORKPATH=$(dirname "$PWD")
export WORKDIR=${WORKPATH}/../../../ export WORKDIR=${WORKPATH}/../../../
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export VLLM_LLM_MODEL_ID="Intel/neural-chat-7b-v3-3" export VLLM_LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
export HF_CACHE_DIR="./data" export HF_CACHE_DIR="./data"
export MODEL_CACHE="./data" export MODEL_CACHE="./data"
@@ -42,7 +42,7 @@ export CRAG_SERVER=http://${ip_address}:${CRAG_SERVER_PORT}
export WORKER_AGENT_URL="http://${ip_address}:${WORKER_RAG_AGENT_PORT}/v1/chat/completions" export WORKER_AGENT_URL="http://${ip_address}:${WORKER_RAG_AGENT_PORT}/v1/chat/completions"
export SQL_AGENT_URL="http://${ip_address}:${WORKER_SQL_AGENT_PORT}/v1/chat/completions" export SQL_AGENT_URL="http://${ip_address}:${WORKER_SQL_AGENT_PORT}/v1/chat/completions"
export HF_CACHE_DIR=${HF_CACHE_DIR} export HF_CACHE_DIR=${HF_CACHE_DIR}
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export no_proxy=${no_proxy} export no_proxy=${no_proxy}
export http_proxy=${http_proxy} export http_proxy=${http_proxy}
export https_proxy=${https_proxy} export https_proxy=${https_proxy}

View File

@@ -33,7 +33,7 @@ fi
# retriever # retriever
export host_ip=$(hostname -I | awk '{print $1}') export host_ip=$(hostname -I | awk '{print $1}')
export HF_CACHE_DIR=${HF_CACHE_DIR} export HF_CACHE_DIR=${HF_CACHE_DIR}
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export no_proxy=${no_proxy} export no_proxy=${no_proxy}
export http_proxy=${http_proxy} export http_proxy=${http_proxy}
export https_proxy=${https_proxy} export https_proxy=${https_proxy}

View File

@@ -16,7 +16,7 @@ services:
with_memory: false with_memory: false
recursion_limit: ${recursion_limit_worker} recursion_limit: ${recursion_limit_worker}
llm_engine: vllm llm_engine: vllm
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
llm_endpoint_url: ${LLM_ENDPOINT_URL} llm_endpoint_url: ${LLM_ENDPOINT_URL}
model: ${LLM_MODEL_ID} model: ${LLM_MODEL_ID}
temperature: ${temperature} temperature: ${temperature}
@@ -50,7 +50,7 @@ services:
use_hints: false use_hints: false
recursion_limit: ${recursion_limit_worker} recursion_limit: ${recursion_limit_worker}
llm_engine: vllm llm_engine: vllm
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
llm_endpoint_url: ${LLM_ENDPOINT_URL} llm_endpoint_url: ${LLM_ENDPOINT_URL}
model: ${LLM_MODEL_ID} model: ${LLM_MODEL_ID}
temperature: ${temperature} temperature: ${temperature}
@@ -79,7 +79,7 @@ services:
with_memory: true with_memory: true
recursion_limit: ${recursion_limit_supervisor} recursion_limit: ${recursion_limit_supervisor}
llm_engine: vllm llm_engine: vllm
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
llm_endpoint_url: ${LLM_ENDPOINT_URL} llm_endpoint_url: ${LLM_ENDPOINT_URL}
model: ${LLM_MODEL_ID} model: ${LLM_MODEL_ID}
temperature: ${temperature} temperature: ${temperature}
@@ -122,7 +122,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
HABANA_VISIBLE_DEVICES: all HABANA_VISIBLE_DEVICES: all
OMPI_MCA_btl_vader_single_copy_mechanism: none OMPI_MCA_btl_vader_single_copy_mechanism: none
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}

View File

@@ -16,8 +16,8 @@ export ip_address=$(hostname -I | awk '{print $1}')
# LLM related environment variables # LLM related environment variables
export HF_CACHE_DIR=${HF_CACHE_DIR} export HF_CACHE_DIR=${HF_CACHE_DIR}
ls $HF_CACHE_DIR ls $HF_CACHE_DIR
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export HF_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HF_TOKEN=${HF_TOKEN}
export LLM_MODEL_ID="meta-llama/Llama-3.3-70B-Instruct" export LLM_MODEL_ID="meta-llama/Llama-3.3-70B-Instruct"
export NUM_SHARDS=4 export NUM_SHARDS=4
export LLM_ENDPOINT_URL="http://${ip_address}:8086" export LLM_ENDPOINT_URL="http://${ip_address}:8086"

View File

@@ -13,7 +13,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
HABANA_VISIBLE_DEVICES: all HABANA_VISIBLE_DEVICES: all

View File

@@ -3,7 +3,7 @@
host_ip=$(hostname -I | awk '{print $1}') host_ip=$(hostname -I | awk '{print $1}')
export HF_CACHE_DIR=${HF_CACHE_DIR} export HF_CACHE_DIR=${HF_CACHE_DIR}
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export no_proxy=${no_proxy} export no_proxy=${no_proxy}
export http_proxy=${http_proxy} export http_proxy=${http_proxy}
export https_proxy=${https_proxy} export https_proxy=${https_proxy}

View File

@@ -20,7 +20,7 @@ function start_retrieval_tool() {
cd $WORKPATH/../DocIndexRetriever/docker_compose/intel/cpu/xeon cd $WORKPATH/../DocIndexRetriever/docker_compose/intel/cpu/xeon
host_ip=$(hostname -I | awk '{print $1}') host_ip=$(hostname -I | awk '{print $1}')
export HF_CACHE_DIR=${HF_CACHE_DIR} export HF_CACHE_DIR=${HF_CACHE_DIR}
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export no_proxy=${no_proxy} export no_proxy=${no_proxy}
export http_proxy=${http_proxy} export http_proxy=${http_proxy}
export https_proxy=${https_proxy} export https_proxy=${https_proxy}

View File

@@ -11,8 +11,8 @@ export ip_address=$(hostname -I | awk '{print $1}')
export host_ip=$ip_address export host_ip=$ip_address
echo "ip_address=${ip_address}" echo "ip_address=${ip_address}"
export TOOLSET_PATH=$WORKPATH/tools/ export TOOLSET_PATH=$WORKPATH/tools/
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
HF_TOKEN=${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN=${HF_TOKEN}
model="meta-llama/Llama-3.3-70B-Instruct" #"meta-llama/Meta-Llama-3.1-70B-Instruct" model="meta-llama/Llama-3.3-70B-Instruct" #"meta-llama/Meta-Llama-3.1-70B-Instruct"
export HF_CACHE_DIR=${model_cache:-"/data2/huggingface"} export HF_CACHE_DIR=${model_cache:-"/data2/huggingface"}

View File

@@ -7,7 +7,7 @@ WORKPATH=$(dirname "$PWD")
export WORKDIR=$WORKPATH/../../ export WORKDIR=$WORKPATH/../../
echo "WORKDIR=${WORKDIR}" echo "WORKDIR=${WORKDIR}"
export ip_address=$(hostname -I | awk '{print $1}') export ip_address=$(hostname -I | awk '{print $1}')
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export TOOLSET_PATH=$WORKDIR/GenAIExamples/AgentQnA/tools/ export TOOLSET_PATH=$WORKDIR/GenAIExamples/AgentQnA/tools/
export no_proxy="$no_proxy,rag-agent-endpoint,sql-agent-endpoint,react-agent-endpoint,agent-ui,vllm-gaudi-server,jaeger,grafana,prometheus,127.0.0.1,localhost,0.0.0.0,$ip_address" export no_proxy="$no_proxy,rag-agent-endpoint,sql-agent-endpoint,react-agent-endpoint,agent-ui,vllm-gaudi-server,jaeger,grafana,prometheus,127.0.0.1,localhost,0.0.0.0,$ip_address"
IMAGE_REPO=${IMAGE_REPO:-"opea"} IMAGE_REPO=${IMAGE_REPO:-"opea"}

View File

@@ -9,7 +9,7 @@ ls $WORKPATH
export WORKDIR=$WORKPATH/../../ export WORKDIR=$WORKPATH/../../
echo "WORKDIR=${WORKDIR}" echo "WORKDIR=${WORKDIR}"
export ip_address=$(hostname -I | awk '{print $1}') export ip_address=$(hostname -I | awk '{print $1}')
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export TOOLSET_PATH=$WORKPATH/tools/ export TOOLSET_PATH=$WORKPATH/tools/
IMAGE_REPO=${IMAGE_REPO:-"opea"} IMAGE_REPO=${IMAGE_REPO:-"opea"}
IMAGE_TAG=${IMAGE_TAG:-"latest"} IMAGE_TAG=${IMAGE_TAG:-"latest"}

View File

@@ -8,7 +8,7 @@ WORKPATH=$(dirname "$PWD")
export WORKDIR=${WORKPATH}/../../ export WORKDIR=${WORKPATH}/../../
echo "WORKDIR=${WORKDIR}" echo "WORKDIR=${WORKDIR}"
export ip_address=$(hostname -I | awk '{print $1}') export ip_address=$(hostname -I | awk '{print $1}')
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export TOOLSET_PATH=$WORKPATH/tools/ export TOOLSET_PATH=$WORKPATH/tools/
IMAGE_REPO=${IMAGE_REPO:-"opea"} IMAGE_REPO=${IMAGE_REPO:-"opea"}
IMAGE_TAG=${IMAGE_TAG:-"latest"} IMAGE_TAG=${IMAGE_TAG:-"latest"}

View File

@@ -40,7 +40,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
command: --model-id ${LLM_MODEL_ID} command: --model-id ${LLM_MODEL_ID}

View File

@@ -35,8 +35,8 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
WILM_USE_TRITON_FLASH_ATTENTION: 0 WILM_USE_TRITON_FLASH_ATTENTION: 0

View File

@@ -7,7 +7,7 @@
# export host_ip=<your External Public IP> # export host_ip=$(hostname -I | awk '{print $1}') # export host_ip=<your External Public IP> # export host_ip=$(hostname -I | awk '{print $1}')
export host_ip=${ip_address} export host_ip=${ip_address}
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
# <token> # <token>
export LLM_MODEL_ID=Intel/neural-chat-7b-v3-3 export LLM_MODEL_ID=Intel/neural-chat-7b-v3-3

View File

@@ -8,7 +8,7 @@
export host_ip=${ip_address} export host_ip=${ip_address}
export external_host_ip=${ip_address} export external_host_ip=${ip_address}
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export HF_CACHE_DIR="./data" export HF_CACHE_DIR="./data"
export LLM_MODEL_ID="Intel/neural-chat-7b-v3-3" export LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
export VLLM_SERVICE_PORT="8081" export VLLM_SERVICE_PORT="8081"

View File

@@ -36,7 +36,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}
VLLM_TORCH_PROFILER_DIR: "/mnt" VLLM_TORCH_PROFILER_DIR: "/mnt"
LLM_SERVER_PORT: ${LLM_SERVER_PORT} LLM_SERVER_PORT: ${LLM_SERVER_PORT}

View File

@@ -40,7 +40,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}
VLLM_TORCH_PROFILER_DIR: "/mnt" VLLM_TORCH_PROFILER_DIR: "/mnt"
LLM_SERVER_PORT: ${LLM_SERVER_PORT} LLM_SERVER_PORT: ${LLM_SERVER_PORT}

View File

@@ -36,7 +36,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
LLM_SERVER_PORT: ${LLM_SERVER_PORT} LLM_SERVER_PORT: ${LLM_SERVER_PORT}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://$host_ip:${LLM_SERVER_PORT}/health || exit 1"] test: ["CMD-SHELL", "curl -f http://$host_ip:${LLM_SERVER_PORT}/health || exit 1"]

View File

@@ -5,7 +5,7 @@
# export host_ip=<your External Public IP> # export host_ip=<your External Public IP>
export host_ip=$(hostname -I | awk '{print $1}') export host_ip=$(hostname -I | awk '{print $1}')
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
# <token> # <token>
export LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct" export LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"

View File

@@ -45,7 +45,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
HABANA_VISIBLE_DEVICES: all HABANA_VISIBLE_DEVICES: all

View File

@@ -45,7 +45,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
HABANA_VISIBLE_DEVICES: all HABANA_VISIBLE_DEVICES: all

View File

@@ -5,7 +5,7 @@
# export host_ip=<your External Public IP> # export host_ip=<your External Public IP>
export host_ip=$(hostname -I | awk '{print $1}') export host_ip=$(hostname -I | awk '{print $1}')
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
# <token> # <token>
export LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct" export LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"

View File

@@ -68,7 +68,7 @@ Then run the command `docker images`, you will have following images ready:
Before starting the services with `docker compose`, you have to recheck the following environment variables. Before starting the services with `docker compose`, you have to recheck the following environment variables.
```bash ```bash
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export host_ip=$(hostname -I | awk '{print $1}') export host_ip=$(hostname -I | awk '{print $1}')
export TGI_SERVICE_PORT=3006 export TGI_SERVICE_PORT=3006

View File

@@ -52,8 +52,8 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
shm_size: 1g shm_size: 1g
devices: devices:
- /dev/kfd:/dev/kfd - /dev/kfd:/dev/kfd

View File

@@ -3,7 +3,7 @@
# Copyright (C) 2024 Intel Corporation # Copyright (C) 2024 Intel Corporation
# SPDX-License-Identifier: Apache-2.0 # SPDX-License-Identifier: Apache-2.0
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export OPENAI_API_KEY=${OPENAI_API_KEY} export OPENAI_API_KEY=${OPENAI_API_KEY}
export host_ip=$(hostname -I | awk '{print $1}') export host_ip=$(hostname -I | awk '{print $1}')

View File

@@ -37,7 +37,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://${host_ip}:3006/health || exit 1"] test: ["CMD-SHELL", "curl -f http://${host_ip}:3006/health || exit 1"]
interval: 10s interval: 10s

View File

@@ -6,7 +6,7 @@ pushd "../../../../../" > /dev/null
source .set_env.sh source .set_env.sh
popd > /dev/null popd > /dev/null
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export host_ip=$(hostname -I | awk '{print $1}') export host_ip=$(hostname -I | awk '{print $1}')
export LLM_MODEL_ID=Intel/neural-chat-7b-v3-3 export LLM_MODEL_ID=Intel/neural-chat-7b-v3-3
export WAV2LIP_ENDPOINT=http://$host_ip:7860 export WAV2LIP_ENDPOINT=http://$host_ip:7860

View File

@@ -48,7 +48,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
HABANA_VISIBLE_DEVICES: all HABANA_VISIBLE_DEVICES: all

View File

@@ -6,7 +6,7 @@ pushd "../../../../../" > /dev/null
source .set_env.sh source .set_env.sh
popd > /dev/null popd > /dev/null
export HUGGINGFACEHUB_API_TOKEN=$HUGGINGFACEHUB_API_TOKEN export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export host_ip=$(hostname -I | awk '{print $1}') export host_ip=$(hostname -I | awk '{print $1}')
export LLM_MODEL_ID=Intel/neural-chat-7b-v3-3 export LLM_MODEL_ID=Intel/neural-chat-7b-v3-3

View File

@@ -6,7 +6,7 @@ export HOST_IP=${ip_address}
export HOST_IP_EXTERNAL=${ip_address} export HOST_IP_EXTERNAL=${ip_address}
export CHATQNA_EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5" export CHATQNA_EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
export CHATQNA_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export CHATQNA_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export CHATQNA_LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct" export CHATQNA_LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
export CHATQNA_RERANK_MODEL_ID="BAAI/bge-reranker-base" export CHATQNA_RERANK_MODEL_ID="BAAI/bge-reranker-base"

View File

@@ -6,7 +6,7 @@ export HOST_IP=${ip_address}
export HOST_IP_EXTERNAL=${ip_address} export HOST_IP_EXTERNAL=${ip_address}
export CHATQNA_EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5" export CHATQNA_EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
export CHATQNA_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export CHATQNA_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export CHATQNA_LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct" export CHATQNA_LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
export CHATQNA_RERANK_MODEL_ID="BAAI/bge-reranker-base" export CHATQNA_RERANK_MODEL_ID="BAAI/bge-reranker-base"

View File

@@ -6,7 +6,7 @@ export HOST_IP=${ip_address}
export HOST_IP_EXTERNAL=${ip_address} export HOST_IP_EXTERNAL=${ip_address}
export CHATQNA_EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5" export CHATQNA_EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
export CHATQNA_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export CHATQNA_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export CHATQNA_LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct" export CHATQNA_LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
export CHATQNA_RERANK_MODEL_ID="BAAI/bge-reranker-base" export CHATQNA_RERANK_MODEL_ID="BAAI/bge-reranker-base"

View File

@@ -6,7 +6,7 @@ export HOST_IP=${ip_address}
export HOST_IP_EXTERNAL=${ip_address} export HOST_IP_EXTERNAL=${ip_address}
export CHATQNA_EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5" export CHATQNA_EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
export CHATQNA_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export CHATQNA_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export CHATQNA_LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct" export CHATQNA_LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
export CHATQNA_RERANK_MODEL_ID="BAAI/bge-reranker-base" export CHATQNA_RERANK_MODEL_ID="BAAI/bge-reranker-base"

View File

@@ -183,7 +183,7 @@ export https_proxy=${your_http_proxy}
export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5" export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
export RERANK_MODEL_ID="BAAI/bge-reranker-base" export RERANK_MODEL_ID="BAAI/bge-reranker-base"
export INDEX_NAME="rag-redis" export INDEX_NAME="rag-redis"
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export OLLAMA_HOST=${host_ip} export OLLAMA_HOST=${host_ip}
export OLLAMA_MODEL="llama3.2" export OLLAMA_MODEL="llama3.2"
``` ```
@@ -194,7 +194,7 @@ export OLLAMA_MODEL="llama3.2"
set EMBEDDING_MODEL_ID=BAAI/bge-base-en-v1.5 set EMBEDDING_MODEL_ID=BAAI/bge-base-en-v1.5
set RERANK_MODEL_ID=BAAI/bge-reranker-base set RERANK_MODEL_ID=BAAI/bge-reranker-base
set INDEX_NAME=rag-redis set INDEX_NAME=rag-redis
set HUGGINGFACEHUB_API_TOKEN=%HUGGINGFACEHUB_API_TOKEN% set HUGGINGFACEHUB_API_TOKEN=%HF_TOKEN%
set OLLAMA_HOST=host.docker.internal set OLLAMA_HOST=host.docker.internal
set OLLAMA_MODEL="llama3.2" set OLLAMA_MODEL="llama3.2"
``` ```

View File

@@ -24,7 +24,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_ENDPOINT: http://tei-embedding-service:80 TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s interval: 10s
@@ -60,7 +60,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80 TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS" RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
restart: unless-stopped restart: unless-stopped
@@ -76,7 +76,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
command: --model-id ${RERANK_MODEL_ID} --auto-truncate command: --model-id ${RERANK_MODEL_ID} --auto-truncate

View File

@@ -9,7 +9,7 @@ popd > /dev/null
export host_ip=$(hostname -I | awk '{print $1}') export host_ip=$(hostname -I | awk '{print $1}')
if [ -z "${HUGGINGFACEHUB_API_TOKEN}" ]; then if [ -z "${HF_TOKEN}" ]; then
echo "Error: HUGGINGFACEHUB_API_TOKEN is not set. Please set HUGGINGFACEHUB_API_TOKEN." echo "Error: HUGGINGFACEHUB_API_TOKEN is not set. Please set HUGGINGFACEHUB_API_TOKEN."
fi fi
@@ -17,7 +17,7 @@ if [ -z "${host_ip}" ]; then
echo "Error: host_ip is not set. Please set host_ip first." echo "Error: host_ip is not set. Please set host_ip first."
fi fi
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5" export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
export RERANK_MODEL_ID="BAAI/bge-reranker-base" export RERANK_MODEL_ID="BAAI/bge-reranker-base"
export INDEX_NAME="rag-redis" export INDEX_NAME="rag-redis"

View File

@@ -31,7 +31,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_ENDPOINT: http://tei-embedding-service:80 TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s interval: 10s
@@ -67,7 +67,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80 TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS" RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
restart: unless-stopped restart: unless-stopped
@@ -83,7 +83,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
command: --model-id ${RERANK_MODEL_ID} --auto-truncate command: --model-id ${RERANK_MODEL_ID} --auto-truncate
@@ -99,7 +99,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}
VLLM_TORCH_PROFILER_DIR: "/mnt" VLLM_TORCH_PROFILER_DIR: "/mnt"
VLLM_CPU_KVCACHE_SPACE: 40 VLLM_CPU_KVCACHE_SPACE: 40

View File

@@ -24,7 +24,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_ENDPOINT: http://tei-embedding-service:80 TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s interval: 10s
@@ -60,7 +60,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80 TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS" RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
restart: unless-stopped restart: unless-stopped
@@ -76,7 +76,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
command: --model-id ${RERANK_MODEL_ID} --auto-truncate command: --model-id ${RERANK_MODEL_ID} --auto-truncate

View File

@@ -24,7 +24,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_ENDPOINT: http://tei-embedding-service:80 TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s interval: 10s
@@ -60,7 +60,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80 TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS" RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
restart: unless-stopped restart: unless-stopped
@@ -76,7 +76,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
command: --model-id ${RERANK_MODEL_ID} --auto-truncate command: --model-id ${RERANK_MODEL_ID} --auto-truncate

View File

@@ -35,7 +35,7 @@ services:
DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_MARIADBVECTOR" DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_MARIADBVECTOR"
MARIADB_CONNECTION_URL: mariadb+mariadbconnector://${MARIADB_USER}:${MARIADB_PASSWORD}@mariadb-server:3306/${MARIADB_DATABASE} MARIADB_CONNECTION_URL: mariadb+mariadbconnector://${MARIADB_USER}:${MARIADB_PASSWORD}@mariadb-server:3306/${MARIADB_DATABASE}
TEI_ENDPOINT: http://tei-embedding-service:80 TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s interval: 10s
@@ -69,7 +69,7 @@ services:
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
MARIADB_CONNECTION_URL: mariadb+mariadbconnector://${MARIADB_USER}:${MARIADB_PASSWORD}@mariadb-server:3306/${MARIADB_DATABASE} MARIADB_CONNECTION_URL: mariadb+mariadbconnector://${MARIADB_USER}:${MARIADB_PASSWORD}@mariadb-server:3306/${MARIADB_DATABASE}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_MARIADBVECTOR" RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_MARIADBVECTOR"
restart: unless-stopped restart: unless-stopped
@@ -85,7 +85,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
command: --model-id ${RERANK_MODEL_ID} --auto-truncate command: --model-id ${RERANK_MODEL_ID} --auto-truncate
@@ -101,7 +101,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}
VLLM_TORCH_PROFILER_DIR: "/mnt" VLLM_TORCH_PROFILER_DIR: "/mnt"
VLLM_CPU_KVCACHE_SPACE: 40 VLLM_CPU_KVCACHE_SPACE: 40

View File

@@ -75,7 +75,7 @@ services:
MILVUS_HOST: ${host_ip} MILVUS_HOST: ${host_ip}
MILVUS_PORT: 19530 MILVUS_PORT: 19530
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80 TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
EMBEDDING_MODEL_ID: ${EMBEDDING_MODEL_ID} EMBEDDING_MODEL_ID: ${EMBEDDING_MODEL_ID}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
healthcheck: healthcheck:
@@ -107,7 +107,7 @@ services:
MILVUS_HOST: ${host_ip} MILVUS_HOST: ${host_ip}
MILVUS_PORT: 19530 MILVUS_PORT: 19530
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80 TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_MILVUS" RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_MILVUS"
restart: unless-stopped restart: unless-stopped
@@ -138,7 +138,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
command: --model-id ${RERANK_MODEL_ID} --auto-truncate command: --model-id ${RERANK_MODEL_ID} --auto-truncate
@@ -155,7 +155,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}
VLLM_TORCH_PROFILER_DIR: "/mnt" VLLM_TORCH_PROFILER_DIR: "/mnt"
healthcheck: healthcheck:

View File

@@ -20,7 +20,7 @@ services:
PINECONE_INDEX_NAME: ${PINECONE_INDEX_NAME} PINECONE_INDEX_NAME: ${PINECONE_INDEX_NAME}
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80 TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_PINECONE" DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_PINECONE"
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
@@ -55,7 +55,7 @@ services:
PINECONE_INDEX_NAME: ${PINECONE_INDEX_NAME} PINECONE_INDEX_NAME: ${PINECONE_INDEX_NAME}
LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY}
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80 TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_PINECONE" RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_PINECONE"
restart: unless-stopped restart: unless-stopped
@@ -71,7 +71,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
command: --model-id ${RERANK_MODEL_ID} --auto-truncate command: --model-id ${RERANK_MODEL_ID} --auto-truncate
@@ -87,7 +87,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}
VLLM_TORCH_PROFILER_DIR: "/mnt" VLLM_TORCH_PROFILER_DIR: "/mnt"
command: --model $LLM_MODEL_ID --host 0.0.0.0 --port 80 command: --model $LLM_MODEL_ID --host 0.0.0.0 --port 80

View File

@@ -24,7 +24,7 @@ services:
QDRANT_PORT: 6333 QDRANT_PORT: 6333
QDRANT_INDEX_NAME: ${INDEX_NAME} QDRANT_INDEX_NAME: ${INDEX_NAME}
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80 TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_QDRANT" DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_QDRANT"
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
@@ -76,7 +76,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
command: --model-id ${RERANK_MODEL_ID} --auto-truncate command: --model-id ${RERANK_MODEL_ID} --auto-truncate
@@ -92,7 +92,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}
VLLM_TORCH_PROFILER_DIR: "/mnt" VLLM_TORCH_PROFILER_DIR: "/mnt"
command: --model $LLM_MODEL_ID --host 0.0.0.0 --port 80 command: --model $LLM_MODEL_ID --host 0.0.0.0 --port 80

View File

@@ -31,7 +31,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_ENDPOINT: http://tei-embedding-service:80 TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
tei-embedding-service: tei-embedding-service:
image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5 image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5
container_name: tei-embedding-server container_name: tei-embedding-server
@@ -61,7 +61,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80 TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS" RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
restart: unless-stopped restart: unless-stopped
@@ -77,7 +77,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
command: --model-id ${RERANK_MODEL_ID} --auto-truncate command: --model-id ${RERANK_MODEL_ID} --auto-truncate

View File

@@ -24,7 +24,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_ENDPOINT: http://tei-embedding-service:80 TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s interval: 10s
@@ -60,7 +60,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80 TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS" RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
restart: unless-stopped restart: unless-stopped
@@ -76,7 +76,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
command: --model-id ${RERANK_MODEL_ID} --auto-truncate command: --model-id ${RERANK_MODEL_ID} --auto-truncate
@@ -92,7 +92,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
command: --model-id ${LLM_MODEL_ID} --cuda-graphs 0 command: --model-id ${LLM_MODEL_ID} --cuda-graphs 0

View File

@@ -24,7 +24,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_ENDPOINT: http://tei-embedding-service:80 TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s interval: 10s
@@ -60,7 +60,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80 TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS" RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
restart: unless-stopped restart: unless-stopped
@@ -76,7 +76,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}
VLLM_TORCH_PROFILER_DIR: "/mnt" VLLM_TORCH_PROFILER_DIR: "/mnt"
command: --model $LLM_MODEL_ID --host 0.0.0.0 --port 80 command: --model $LLM_MODEL_ID --host 0.0.0.0 --port 80

View File

@@ -7,7 +7,7 @@ pushd "../../../../../" > /dev/null
source .set_env.sh source .set_env.sh
popd > /dev/null popd > /dev/null
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export HF_TOKEN=${HF_TOKEN} export HF_TOKEN=${HF_TOKEN}
export host_ip=${ip_address} export host_ip=${ip_address}
export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5" export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"

View File

@@ -7,7 +7,7 @@ pushd "../../../../../" > /dev/null
source .set_env.sh source .set_env.sh
popd > /dev/null popd > /dev/null
if [ -z "${HUGGINGFACEHUB_API_TOKEN}" ]; then if [ -z "${HF_TOKEN}" ]; then
echo "Error: HUGGINGFACEHUB_API_TOKEN is not set. Please set HUGGINGFACEHUB_API_TOKEN." echo "Error: HUGGINGFACEHUB_API_TOKEN is not set. Please set HUGGINGFACEHUB_API_TOKEN."
fi fi
@@ -15,7 +15,7 @@ export host_ip=$(hostname -I | awk '{print $1}')
export MARIADB_DATABASE="vectordb" export MARIADB_DATABASE="vectordb"
export MARIADB_USER="chatqna" export MARIADB_USER="chatqna"
export MARIADB_PASSWORD="password" export MARIADB_PASSWORD="password"
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5" export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
export RERANK_MODEL_ID="BAAI/bge-reranker-base" export RERANK_MODEL_ID="BAAI/bge-reranker-base"
export LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct" export LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"

View File

@@ -31,7 +31,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_ENDPOINT: http://tei-embedding-service:80 TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s interval: 10s
@@ -67,7 +67,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80 TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
restart: unless-stopped restart: unless-stopped
tei-reranking-service: tei-reranking-service:
image: ghcr.io/huggingface/tei-gaudi:1.5.0 image: ghcr.io/huggingface/tei-gaudi:1.5.0
@@ -101,7 +101,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
HABANA_VISIBLE_DEVICES: all HABANA_VISIBLE_DEVICES: all
OMPI_MCA_btl_vader_single_copy_mechanism: none OMPI_MCA_btl_vader_single_copy_mechanism: none
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}

View File

@@ -24,7 +24,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_ENDPOINT: http://tei-embedding-service:80 TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
@@ -61,7 +61,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80 TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
restart: unless-stopped restart: unless-stopped
tei-reranking-service: tei-reranking-service:
image: ghcr.io/huggingface/tei-gaudi:1.5.0 image: ghcr.io/huggingface/tei-gaudi:1.5.0

View File

@@ -24,7 +24,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_ENDPOINT: http://tei-embedding-service:80 TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
@@ -61,7 +61,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80 TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
restart: unless-stopped restart: unless-stopped
tei-reranking-service: tei-reranking-service:
image: ghcr.io/huggingface/tei-gaudi:1.5.0 image: ghcr.io/huggingface/tei-gaudi:1.5.0
@@ -95,7 +95,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
HABANA_VISIBLE_DEVICES: all HABANA_VISIBLE_DEVICES: all

View File

@@ -24,7 +24,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_ENDPOINT: http://tei-embedding-service:80 TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s interval: 10s
@@ -42,7 +42,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
HABANA_VISIBLE_DEVICES: all HABANA_VISIBLE_DEVICES: all
OMPI_MCA_btl_vader_single_copy_mechanism: none OMPI_MCA_btl_vader_single_copy_mechanism: none
GURADRAILS_MODEL_ID: ${GURADRAILS_MODEL_ID} GURADRAILS_MODEL_ID: ${GURADRAILS_MODEL_ID}
@@ -73,7 +73,7 @@ services:
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
SAFETY_GUARD_MODEL_ID: ${GURADRAILS_MODEL_ID} SAFETY_GUARD_MODEL_ID: ${GURADRAILS_MODEL_ID}
SAFETY_GUARD_ENDPOINT: http://vllm-guardrails-service:80 SAFETY_GUARD_ENDPOINT: http://vllm-guardrails-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
restart: unless-stopped restart: unless-stopped
tei-embedding-service: tei-embedding-service:
image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5 image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5
@@ -104,7 +104,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80 TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS" RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
restart: unless-stopped restart: unless-stopped
@@ -140,7 +140,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
HABANA_VISIBLE_DEVICES: all HABANA_VISIBLE_DEVICES: all
OMPI_MCA_btl_vader_single_copy_mechanism: none OMPI_MCA_btl_vader_single_copy_mechanism: none
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}

View File

@@ -24,7 +24,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_ENDPOINT: http://tei-embedding-service:80 TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s interval: 10s
@@ -60,7 +60,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80 TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS" RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
restart: unless-stopped restart: unless-stopped
@@ -96,7 +96,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
HABANA_VISIBLE_DEVICES: all HABANA_VISIBLE_DEVICES: all

View File

@@ -24,7 +24,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_ENDPOINT: http://tei-embedding-service:80 TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s interval: 10s
@@ -60,7 +60,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80 TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS" RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
restart: unless-stopped restart: unless-stopped
@@ -75,7 +75,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
HABANA_VISIBLE_DEVICES: all HABANA_VISIBLE_DEVICES: all
OMPI_MCA_btl_vader_single_copy_mechanism: none OMPI_MCA_btl_vader_single_copy_mechanism: none
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}

View File

@@ -123,7 +123,7 @@ View the docker input parameters in `./ChatQnA/docker_compose/intel/hpu/gaudi/co
environment: environment:
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN}
HABANA_VISIBLE_DEVICES: all HABANA_VISIBLE_DEVICES: all
OMPI_MCA_btl_vader_single_copy_mechanism: none OMPI_MCA_btl_vader_single_copy_mechanism: none
ENABLE_HPU_GRAPH: true ENABLE_HPU_GRAPH: true

View File

@@ -92,7 +92,7 @@ cat <<EOF > .env
# Set all required ENV values # Set all required ENV values
export TAG=${TAG} export TAG=${TAG}
export EMBEDDING_MODEL_ID=${EMBEDDING_MODEL_ID} export EMBEDDING_MODEL_ID=${EMBEDDING_MODEL_ID}
export HUGGINGFACEHUB_API_TOKEN=$HUGGINGFACEHUB_API_TOKEN export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export RERANK_MODEL_ID=${RERANK_MODEL_ID} export RERANK_MODEL_ID=${RERANK_MODEL_ID}
export LLM_MODEL_ID=${LLM_MODEL_ID} export LLM_MODEL_ID=${LLM_MODEL_ID}
export INDEX_NAME=${INDEX_NAME} export INDEX_NAME=${INDEX_NAME}

View File

@@ -7,7 +7,7 @@ pushd "../../../../../" > /dev/null
source .set_env.sh source .set_env.sh
popd > /dev/null popd > /dev/null
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export HF_TOKEN=${HF_TOKEN} export HF_TOKEN=${HF_TOKEN}
export host_ip=${ip_address} export host_ip=${ip_address}
export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5" export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"

View File

@@ -24,7 +24,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_ENDPOINT: http://tei-embedding-service:80 TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s interval: 10s
@@ -76,7 +76,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
command: --model-id ${RERANK_MODEL_ID} --auto-truncate command: --model-id ${RERANK_MODEL_ID} --auto-truncate
@@ -98,7 +98,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
ipc: host ipc: host

View File

@@ -12,7 +12,7 @@ export EXTERNAL_HOST_IP=${ip_address}
export CODEGEN_TGI_SERVICE_PORT=8028 export CODEGEN_TGI_SERVICE_PORT=8028
### A token for accessing repositories with models ### A token for accessing repositories with models
export CODEGEN_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export CODEGEN_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
### Model ID ### Model ID
export CODEGEN_LLM_MODEL_ID="Qwen/Qwen2.5-Coder-7B-Instruct" export CODEGEN_LLM_MODEL_ID="Qwen/Qwen2.5-Coder-7B-Instruct"

View File

@@ -13,7 +13,7 @@ export CODEGEN_VLLM_SERVICE_PORT=8028
export CODEGEN_VLLM_ENDPOINT="http://${HOST_IP}:${CODEGEN_VLLM_SERVICE_PORT}" export CODEGEN_VLLM_ENDPOINT="http://${HOST_IP}:${CODEGEN_VLLM_SERVICE_PORT}"
### A token for accessing repositories with models ### A token for accessing repositories with models
export CODEGEN_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export CODEGEN_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
### Model ID ### Model ID
export CODEGEN_LLM_MODEL_ID="Qwen/Qwen2.5-Coder-7B-Instruct" export CODEGEN_LLM_MODEL_ID="Qwen/Qwen2.5-Coder-7B-Instruct"

View File

@@ -17,7 +17,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
host_ip: ${host_ip} host_ip: ${host_ip}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"]
@@ -39,7 +39,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
host_ip: ${host_ip} host_ip: ${host_ip}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"]
@@ -56,7 +56,7 @@ services:
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
LLM_ENDPOINT: ${LLM_ENDPOINT} LLM_ENDPOINT: ${LLM_ENDPOINT}
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
restart: unless-stopped restart: unless-stopped
llm-tgi-service: llm-tgi-service:
extends: llm-base extends: llm-base
@@ -140,7 +140,7 @@ services:
REDIS_URL: ${REDIS_URL} REDIS_URL: ${REDIS_URL}
REDIS_HOST: ${host_ip} REDIS_HOST: ${host_ip}
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: true LOGFLAG: true
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
@@ -162,7 +162,7 @@ services:
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
host_ip: ${host_ip} host_ip: ${host_ip}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
healthcheck: healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:80/health"] test: ["CMD", "curl", "-f", "http://localhost:80/health"]
interval: 10s interval: 10s
@@ -202,7 +202,7 @@ services:
REDIS_RETRIEVER_PORT: ${REDIS_RETRIEVER_PORT} REDIS_RETRIEVER_PORT: ${REDIS_RETRIEVER_PORT}
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: ${RETRIEVER_COMPONENT_NAME:-OPEA_RETRIEVER_REDIS} RETRIEVER_COMPONENT_NAME: ${RETRIEVER_COMPONENT_NAME:-OPEA_RETRIEVER_REDIS}
restart: unless-stopped restart: unless-stopped

View File

@@ -59,7 +59,7 @@ services:
REDIS_URL: ${REDIS_URL} REDIS_URL: ${REDIS_URL}
REDIS_HOST: ${host_ip} REDIS_HOST: ${host_ip}
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: true LOGFLAG: true
restart: unless-stopped restart: unless-stopped
tei-embedding-serving: tei-embedding-serving:
@@ -76,7 +76,7 @@ services:
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
host_ip: ${host_ip} host_ip: ${host_ip}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
healthcheck: healthcheck:
test: ["CMD", "curl", "-f", "http://${host_ip}:${TEI_EMBEDDER_PORT}/health"] test: ["CMD", "curl", "-f", "http://${host_ip}:${TEI_EMBEDDER_PORT}/health"]
interval: 10s interval: 10s
@@ -116,7 +116,7 @@ services:
REDIS_RETRIEVER_PORT: ${REDIS_RETRIEVER_PORT} REDIS_RETRIEVER_PORT: ${REDIS_RETRIEVER_PORT}
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: ${RETRIEVER_COMPONENT_NAME:-OPEA_RETRIEVER_REDIS} RETRIEVER_COMPONENT_NAME: ${RETRIEVER_COMPONENT_NAME:-OPEA_RETRIEVER_REDIS}
restart: unless-stopped restart: unless-stopped

View File

@@ -17,7 +17,7 @@ services:
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HABANA_VISIBLE_DEVICES: all HABANA_VISIBLE_DEVICES: all
OMPI_MCA_btl_vader_single_copy_mechanism: none OMPI_MCA_btl_vader_single_copy_mechanism: none
HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN}
ENABLE_HPU_GRAPH: true ENABLE_HPU_GRAPH: true
LIMIT_HPU_GRAPH: true LIMIT_HPU_GRAPH: true
USE_FLASH_ATTENTION: true USE_FLASH_ATTENTION: true
@@ -46,7 +46,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
HABANA_VISIBLE_DEVICES: all HABANA_VISIBLE_DEVICES: all
OMPI_MCA_btl_vader_single_copy_mechanism: none OMPI_MCA_btl_vader_single_copy_mechanism: none
VLLM_SKIP_WARMUP: ${VLLM_SKIP_WARMUP:-false} VLLM_SKIP_WARMUP: ${VLLM_SKIP_WARMUP:-false}
@@ -71,7 +71,7 @@ services:
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
LLM_ENDPOINT: ${LLM_ENDPOINT} LLM_ENDPOINT: ${LLM_ENDPOINT}
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
restart: unless-stopped restart: unless-stopped
llm-tgi-service: llm-tgi-service:
extends: llm-base extends: llm-base
@@ -156,7 +156,7 @@ services:
REDIS_URL: ${REDIS_URL} REDIS_URL: ${REDIS_URL}
REDIS_HOST: ${host_ip} REDIS_HOST: ${host_ip}
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: true LOGFLAG: true
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
@@ -178,7 +178,7 @@ services:
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
host_ip: ${host_ip} host_ip: ${host_ip}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
healthcheck: healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:80/health"] test: ["CMD", "curl", "-f", "http://localhost:80/health"]
interval: 10s interval: 10s
@@ -218,7 +218,7 @@ services:
REDIS_RETRIEVER_PORT: ${REDIS_RETRIEVER_PORT} REDIS_RETRIEVER_PORT: ${REDIS_RETRIEVER_PORT}
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: ${RETRIEVER_COMPONENT_NAME:-OPEA_RETRIEVER_REDIS} RETRIEVER_COMPONENT_NAME: ${RETRIEVER_COMPONENT_NAME:-OPEA_RETRIEVER_REDIS}
restart: unless-stopped restart: unless-stopped

View File

@@ -7,8 +7,8 @@ source .set_env.sh
popd > /dev/null popd > /dev/null
export HOST_IP=$(hostname -I | awk '{print $1}') export HOST_IP=$(hostname -I | awk '{print $1}')
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
if [ -z "${HUGGINGFACEHUB_API_TOKEN}" ]; then if [ -z "${HF_TOKEN}" ]; then
echo "Error: HUGGINGFACEHUB_API_TOKEN is not set. Please set HUGGINGFACEHUB_API_TOKEN" echo "Error: HUGGINGFACEHUB_API_TOKEN is not set. Please set HUGGINGFACEHUB_API_TOKEN"
fi fi

View File

@@ -21,7 +21,7 @@ export CODETRANS_TGI_SERVICE_PORT=8008
export CODETRANS_TGI_LLM_ENDPOINT="http://${HOST_IP}:${CODETRANS_TGI_SERVICE_PORT}" export CODETRANS_TGI_LLM_ENDPOINT="http://${HOST_IP}:${CODETRANS_TGI_SERVICE_PORT}"
### A token for accessing repositories with models ### A token for accessing repositories with models
export CODETRANS_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export CODETRANS_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
### The port of the LLM service. On this port, the LLM service will accept connections ### The port of the LLM service. On this port, the LLM service will accept connections
export CODETRANS_LLM_SERVICE_PORT=9000 export CODETRANS_LLM_SERVICE_PORT=9000

View File

@@ -21,7 +21,7 @@ export CODETRANS_VLLM_SERVICE_PORT=8008
export CODETRANS_LLM_ENDPOINT="http://${HOST_IP}:${CODETRANS_VLLM_SERVICE_PORT}" export CODETRANS_LLM_ENDPOINT="http://${HOST_IP}:${CODETRANS_VLLM_SERVICE_PORT}"
### A token for accessing repositories with models ### A token for accessing repositories with models
export CODETRANS_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export CODETRANS_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
### The port of the LLM service. On this port, the LLM service will accept connections ### The port of the LLM service. On this port, the LLM service will accept connections
export CODETRANS_LLM_SERVICE_PORT=9000 export CODETRANS_LLM_SERVICE_PORT=9000

View File

@@ -14,7 +14,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}
VLLM_TORCH_PROFILER_DIR: "/mnt" VLLM_TORCH_PROFILER_DIR: "/mnt"
healthcheck: healthcheck:
@@ -39,7 +39,7 @@ services:
LLM_ENDPOINT: ${LLM_ENDPOINT} LLM_ENDPOINT: ${LLM_ENDPOINT}
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}
LLM_COMPONENT_NAME: ${LLM_COMPONENT_NAME} LLM_COMPONENT_NAME: ${LLM_COMPONENT_NAME}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
restart: unless-stopped restart: unless-stopped
codetrans-xeon-backend-server: codetrans-xeon-backend-server:
image: ${REGISTRY:-opea}/codetrans:${TAG:-latest} image: ${REGISTRY:-opea}/codetrans:${TAG:-latest}

View File

@@ -14,7 +14,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
host_ip: ${host_ip} host_ip: ${host_ip}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://$host_ip:8008/health || exit 1"] test: ["CMD-SHELL", "curl -f http://$host_ip:8008/health || exit 1"]
@@ -38,7 +38,7 @@ services:
LLM_ENDPOINT: ${LLM_ENDPOINT} LLM_ENDPOINT: ${LLM_ENDPOINT}
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}
LLM_COMPONENT_NAME: ${LLM_COMPONENT_NAME} LLM_COMPONENT_NAME: ${LLM_COMPONENT_NAME}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
restart: unless-stopped restart: unless-stopped
codetrans-xeon-backend-server: codetrans-xeon-backend-server:
image: ${REGISTRY:-opea}/codetrans:${TAG:-latest} image: ${REGISTRY:-opea}/codetrans:${TAG:-latest}

View File

@@ -13,7 +13,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
HABANA_VISIBLE_DEVICES: all HABANA_VISIBLE_DEVICES: all
OMPI_MCA_btl_vader_single_copy_mechanism: none OMPI_MCA_btl_vader_single_copy_mechanism: none
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}
@@ -45,7 +45,7 @@ services:
LLM_ENDPOINT: ${LLM_ENDPOINT} LLM_ENDPOINT: ${LLM_ENDPOINT}
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}
LLM_COMPONENT_NAME: ${LLM_COMPONENT_NAME} LLM_COMPONENT_NAME: ${LLM_COMPONENT_NAME}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
restart: unless-stopped restart: unless-stopped
codetrans-gaudi-backend-server: codetrans-gaudi-backend-server:
image: ${REGISTRY:-opea}/codetrans:${TAG:-latest} image: ${REGISTRY:-opea}/codetrans:${TAG:-latest}

View File

@@ -13,7 +13,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
HABANA_VISIBLE_DEVICES: all HABANA_VISIBLE_DEVICES: all
@@ -42,7 +42,7 @@ services:
LLM_ENDPOINT: ${LLM_ENDPOINT} LLM_ENDPOINT: ${LLM_ENDPOINT}
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}
LLM_COMPONENT_NAME: ${LLM_COMPONENT_NAME} LLM_COMPONENT_NAME: ${LLM_COMPONENT_NAME}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
restart: unless-stopped restart: unless-stopped
codetrans-gaudi-backend-server: codetrans-gaudi-backend-server:
image: ${REGISTRY:-opea}/codetrans:${TAG:-latest} image: ${REGISTRY:-opea}/codetrans:${TAG:-latest}

View File

@@ -38,7 +38,7 @@ function build_docker_images() {
function start_services() { function start_services() {
cd $WORKPATH/docker_compose/intel cd $WORKPATH/docker_compose/intel
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export NGINX_PORT=80 export NGINX_PORT=80
export host_ip=${ip_address} export host_ip=${ip_address}
source set_env.sh source set_env.sh

View File

@@ -40,7 +40,7 @@ function build_docker_images() {
function start_services() { function start_services() {
cd $WORKPATH/docker_compose/intel cd $WORKPATH/docker_compose/intel
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export NGINX_PORT=80 export NGINX_PORT=80
export host_ip=${ip_address} export host_ip=${ip_address}

View File

@@ -35,7 +35,7 @@ function build_docker_images() {
function start_services() { function start_services() {
cd $WORKPATH/docker_compose/intel cd $WORKPATH/docker_compose/intel
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export NGINX_PORT=80 export NGINX_PORT=80
export host_ip=${ip_address} export host_ip=${ip_address}

View File

@@ -35,7 +35,7 @@ function build_docker_images() {
function start_services() { function start_services() {
cd $WORKPATH/docker_compose/intel cd $WORKPATH/docker_compose/intel
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export NGINX_PORT=80 export NGINX_PORT=80
export host_ip=${ip_address} export host_ip=${ip_address}

View File

@@ -9,7 +9,7 @@ source .set_env.sh
popd > /dev/null popd > /dev/null
export host_ip=${ip_address} export host_ip=${ip_address}
export DBQNA_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export DBQNA_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export DBQNA_TGI_SERVICE_PORT=8008 export DBQNA_TGI_SERVICE_PORT=8008
export DBQNA_TGI_LLM_ENDPOINT="http://${host_ip}:${DBQNA_TGI_SERVICE_PORT}" export DBQNA_TGI_LLM_ENDPOINT="http://${host_ip}:${DBQNA_TGI_SERVICE_PORT}"
export DBQNA_LLM_MODEL_ID="mistralai/Mistral-7B-Instruct-v0.3" export DBQNA_LLM_MODEL_ID="mistralai/Mistral-7B-Instruct-v0.3"

View File

@@ -61,7 +61,7 @@ export https_proxy=${https_proxy}
export TGI_PORT=8008 export TGI_PORT=8008
export TGI_LLM_ENDPOINT=http://${host_ip}:${TGI_PORT} export TGI_LLM_ENDPOINT=http://${host_ip}:${TGI_PORT}
export HF_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HF_TOKEN=${HF_TOKEN}
export LLM_MODEL_ID="mistralai/Mistral-7B-Instruct-v0.3" export LLM_MODEL_ID="mistralai/Mistral-7B-Instruct-v0.3"
export POSTGRES_USER=postgres export POSTGRES_USER=postgres
export POSTGRES_PASSWORD=testpwd export POSTGRES_PASSWORD=testpwd
@@ -108,7 +108,7 @@ docker run --name test-text2sql-postgres --ipc=host -e POSTGRES_USER=${POSTGRES_
```bash ```bash
docker run -d --name="test-text2sql-tgi-endpoint" --ipc=host -p $TGI_PORT:80 -v ./data:/data --shm-size 1g -e HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} -e HF_TOKEN=${HF_TOKEN} -e model=${model} ghcr.io/huggingface/text-generation-inference:2.4.1 --model-id $model docker run -d --name="test-text2sql-tgi-endpoint" --ipc=host -p $TGI_PORT:80 -v ./data:/data --shm-size 1g -e HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN} -e HF_TOKEN=${HF_TOKEN} -e model=${model} ghcr.io/huggingface/text-generation-inference:2.4.1 --model-id $model
``` ```
- Start Text-to-SQL Service - Start Text-to-SQL Service

View File

@@ -13,8 +13,8 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
shm_size: 1g shm_size: 1g
command: --model-id ${LLM_MODEL_ID} command: --model-id ${LLM_MODEL_ID}

View File

@@ -9,8 +9,8 @@ popd > /dev/null
export host_ip=${ip_address} export host_ip=${ip_address}
export no_proxy=$no_proxy,$host_ip,dbqna-xeon-react-ui-server,text2sql-service,tgi-service,postgres-container export no_proxy=$no_proxy,$host_ip,dbqna-xeon-react-ui-server,text2sql-service,tgi-service,postgres-container
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export HF_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HF_TOKEN=${HF_TOKEN}
export POSTGRES_USER=postgres export POSTGRES_USER=postgres
export POSTGRES_PASSWORD=testpwd export POSTGRES_PASSWORD=testpwd
export POSTGRES_DB=chinook export POSTGRES_DB=chinook

View File

@@ -28,7 +28,7 @@ services:
REDIS_HOST: ${REDIS_HOST} REDIS_HOST: ${REDIS_HOST}
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
@@ -49,7 +49,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
host_ip: ${host_ip} host_ip: ${host_ip}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"]
@@ -69,7 +69,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
restart: unless-stopped restart: unless-stopped
@@ -87,7 +87,7 @@ services:
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
REDIS_URL: ${REDIS_URL} REDIS_URL: ${REDIS_URL}
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS" RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
@@ -105,7 +105,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
host_ip: ${host_ip} host_ip: ${host_ip}
@@ -129,7 +129,7 @@ services:
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
RERANK_TYPE: ${RERANK_TYPE} RERANK_TYPE: ${RERANK_TYPE}
TEI_RERANKING_ENDPOINT: ${TEI_RERANKING_ENDPOINT} TEI_RERANKING_ENDPOINT: ${TEI_RERANKING_ENDPOINT}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}

View File

@@ -76,7 +76,7 @@ services:
DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_MILVUS" DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_MILVUS"
MILVUS_HOST: ${MILVUS_HOST} MILVUS_HOST: ${MILVUS_HOST}
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
@@ -107,7 +107,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
host_ip: ${host_ip} host_ip: ${host_ip}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"]
@@ -130,7 +130,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
restart: unless-stopped restart: unless-stopped
@@ -148,7 +148,7 @@ services:
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
MILVUS_HOST: ${host_ip} MILVUS_HOST: ${host_ip}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_MILVUS" RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_MILVUS"
@@ -167,7 +167,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
host_ip: ${host_ip} host_ip: ${host_ip}
@@ -194,7 +194,7 @@ services:
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
RERANK_TYPE: ${RERANK_TYPE} RERANK_TYPE: ${RERANK_TYPE}
TEI_RERANKING_ENDPOINT: ${TEI_RERANKING_ENDPOINT} TEI_RERANKING_ENDPOINT: ${TEI_RERANKING_ENDPOINT}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}

View File

@@ -25,7 +25,7 @@ services:
REDIS_HOST: redis-vector-db REDIS_HOST: redis-vector-db
INDEX_NAME: ${INDEX_NAME:-rag-redis} INDEX_NAME: ${INDEX_NAME:-rag-redis}
TEI_ENDPOINT: http://tei-embedding-service:80 TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
@@ -46,7 +46,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
host_ip: ${host_ip} host_ip: ${host_ip}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"]
@@ -66,7 +66,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80 TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
restart: unless-stopped restart: unless-stopped
@@ -84,7 +84,7 @@ services:
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
REDIS_URL: redis://redis-vector-db:6379 REDIS_URL: redis://redis-vector-db:6379
INDEX_NAME: ${INDEX_NAME:-rag-redis} INDEX_NAME: ${INDEX_NAME:-rag-redis}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80 TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS" RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"

View File

@@ -12,7 +12,7 @@ export RERANK_MODEL_ID="BAAI/bge-reranker-base"
export TEI_EMBEDDING_ENDPOINT="http://${ip_address}:6006" export TEI_EMBEDDING_ENDPOINT="http://${ip_address}:6006"
export TEI_RERANKING_ENDPOINT="http://${ip_address}:8808" export TEI_RERANKING_ENDPOINT="http://${ip_address}:8808"
export TGI_LLM_ENDPOINT="http://${ip_address}:8008" export TGI_LLM_ENDPOINT="http://${ip_address}:8008"
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export MEGA_SERVICE_HOST_IP=${ip_address} export MEGA_SERVICE_HOST_IP=${ip_address}
export EMBEDDING_SERVICE_HOST_IP=${ip_address} export EMBEDDING_SERVICE_HOST_IP=${ip_address}
export RETRIEVER_SERVICE_HOST_IP=${ip_address} export RETRIEVER_SERVICE_HOST_IP=${ip_address}

View File

@@ -28,7 +28,7 @@ services:
REDIS_URL: ${REDIS_URL} REDIS_URL: ${REDIS_URL}
INDEX_NAME: ${INDEX_NAME} INDEX_NAME: ${INDEX_NAME}
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s interval: 10s
@@ -76,7 +76,7 @@ services:
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
restart: unless-stopped restart: unless-stopped
retriever: retriever:
@@ -96,7 +96,7 @@ services:
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS" RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
restart: unless-stopped restart: unless-stopped
tei-reranking-service: tei-reranking-service:
image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.6 image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.6
@@ -111,7 +111,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
host_ip: ${host_ip} host_ip: ${host_ip}
@@ -135,7 +135,7 @@ services:
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
RERANK_TYPE: ${RERANK_TYPE} RERANK_TYPE: ${RERANK_TYPE}
TEI_RERANKING_ENDPOINT: ${TEI_RERANKING_ENDPOINT} TEI_RERANKING_ENDPOINT: ${TEI_RERANKING_ENDPOINT}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}

View File

@@ -76,7 +76,7 @@ services:
DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_MILVUS" DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_MILVUS"
MILVUS_HOST: ${MILVUS_HOST} MILVUS_HOST: ${MILVUS_HOST}
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
@@ -136,7 +136,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
restart: unless-stopped restart: unless-stopped
@@ -154,7 +154,7 @@ services:
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
MILVUS_HOST: ${host_ip} MILVUS_HOST: ${host_ip}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_MILVUS" RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_MILVUS"
@@ -173,7 +173,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
host_ip: ${host_ip} host_ip: ${host_ip}
@@ -200,7 +200,7 @@ services:
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
RERANK_TYPE: ${RERANK_TYPE} RERANK_TYPE: ${RERANK_TYPE}
TEI_RERANKING_ENDPOINT: ${TEI_RERANKING_ENDPOINT} TEI_RERANKING_ENDPOINT: ${TEI_RERANKING_ENDPOINT}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
LOGFLAG: ${LOGFLAG} LOGFLAG: ${LOGFLAG}

View File

@@ -11,7 +11,7 @@ export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
export RERANK_MODEL_ID="BAAI/bge-reranker-base" export RERANK_MODEL_ID="BAAI/bge-reranker-base"
export TEI_EMBEDDING_ENDPOINT="http://${ip_address}:8090" export TEI_EMBEDDING_ENDPOINT="http://${ip_address}:8090"
export TEI_RERANKING_ENDPOINT="http://${ip_address}:8808" export TEI_RERANKING_ENDPOINT="http://${ip_address}:8808"
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export MEGA_SERVICE_HOST_IP=${ip_address} export MEGA_SERVICE_HOST_IP=${ip_address}
export EMBEDDING_SERVICE_HOST_IP=${ip_address} export EMBEDDING_SERVICE_HOST_IP=${ip_address}
export RETRIEVER_SERVICE_HOST_IP=${ip_address} export RETRIEVER_SERVICE_HOST_IP=${ip_address}

View File

@@ -43,7 +43,7 @@ function start_services() {
export TEI_EMBEDDING_ENDPOINT="http://${ip_address}:6006" export TEI_EMBEDDING_ENDPOINT="http://${ip_address}:6006"
export REDIS_URL="redis://${ip_address}:6379" export REDIS_URL="redis://${ip_address}:6379"
export INDEX_NAME="rag-redis" export INDEX_NAME="rag-redis"
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export MEGA_SERVICE_HOST_IP=${ip_address} export MEGA_SERVICE_HOST_IP=${ip_address}
export EMBEDDING_SERVICE_HOST_IP=${ip_address} export EMBEDDING_SERVICE_HOST_IP=${ip_address}
export RETRIEVER_SERVICE_HOST_IP=${ip_address} export RETRIEVER_SERVICE_HOST_IP=${ip_address}

View File

@@ -9,7 +9,7 @@ export DOCSUM_MAX_TOTAL_TOKENS="4096"
export DOCSUM_LLM_MODEL_ID="Intel/neural-chat-7b-v3-3" export DOCSUM_LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
export DOCSUM_TGI_SERVICE_PORT="8008" export DOCSUM_TGI_SERVICE_PORT="8008"
export DOCSUM_TGI_LLM_ENDPOINT="http://${HOST_IP}:${DOCSUM_TGI_SERVICE_PORT}" export DOCSUM_TGI_LLM_ENDPOINT="http://${HOST_IP}:${DOCSUM_TGI_SERVICE_PORT}"
export DOCSUM_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export DOCSUM_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export DOCSUM_WHISPER_PORT="7066" export DOCSUM_WHISPER_PORT="7066"
export ASR_SERVICE_HOST_IP="${HOST_IP}" export ASR_SERVICE_HOST_IP="${HOST_IP}"
export DOCSUM_LLM_SERVER_PORT="9000" export DOCSUM_LLM_SERVER_PORT="9000"

View File

@@ -4,7 +4,7 @@
# SPDX-License-Identifier: Apache-2.0 # SPDX-License-Identifier: Apache-2.0
export HOST_IP=${ip_address} export HOST_IP=${ip_address}
export DOCSUM_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export DOCSUM_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export DOCSUM_MAX_INPUT_TOKENS=2048 export DOCSUM_MAX_INPUT_TOKENS=2048
export DOCSUM_MAX_TOTAL_TOKENS=4096 export DOCSUM_MAX_TOTAL_TOKENS=4096
export DOCSUM_LLM_MODEL_ID="Intel/neural-chat-7b-v3-3" export DOCSUM_LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"

View File

@@ -14,7 +14,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}
VLLM_TORCH_PROFILER_DIR: "/mnt" VLLM_TORCH_PROFILER_DIR: "/mnt"
healthcheck: healthcheck:
@@ -39,7 +39,7 @@ services:
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
LLM_ENDPOINT: ${LLM_ENDPOINT} LLM_ENDPOINT: ${LLM_ENDPOINT}
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
MAX_INPUT_TOKENS: ${MAX_INPUT_TOKENS} MAX_INPUT_TOKENS: ${MAX_INPUT_TOKENS}
MAX_TOTAL_TOKENS: ${MAX_TOTAL_TOKENS} MAX_TOTAL_TOKENS: ${MAX_TOTAL_TOKENS}
DocSum_COMPONENT_NAME: ${DocSum_COMPONENT_NAME} DocSum_COMPONENT_NAME: ${DocSum_COMPONENT_NAME}

View File

@@ -14,7 +14,7 @@ services:
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
TGI_LLM_ENDPOINT: ${TGI_LLM_ENDPOINT} TGI_LLM_ENDPOINT: ${TGI_LLM_ENDPOINT}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
host_ip: ${host_ip} host_ip: ${host_ip}
healthcheck: healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"] test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"]
@@ -39,7 +39,7 @@ services:
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
LLM_ENDPOINT: ${LLM_ENDPOINT} LLM_ENDPOINT: ${LLM_ENDPOINT}
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
MAX_INPUT_TOKENS: ${MAX_INPUT_TOKENS} MAX_INPUT_TOKENS: ${MAX_INPUT_TOKENS}
MAX_TOTAL_TOKENS: ${MAX_TOTAL_TOKENS} MAX_TOTAL_TOKENS: ${MAX_TOTAL_TOKENS}
DocSum_COMPONENT_NAME: ${DocSum_COMPONENT_NAME} DocSum_COMPONENT_NAME: ${DocSum_COMPONENT_NAME}

View File

@@ -13,7 +13,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_TOKEN: ${HF_TOKEN}
HABANA_VISIBLE_DEVICES: all HABANA_VISIBLE_DEVICES: all
OMPI_MCA_btl_vader_single_copy_mechanism: none OMPI_MCA_btl_vader_single_copy_mechanism: none
LLM_MODEL_ID: ${LLM_MODEL_ID} LLM_MODEL_ID: ${LLM_MODEL_ID}
@@ -44,7 +44,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
MAX_INPUT_TOKENS: ${MAX_INPUT_TOKENS} MAX_INPUT_TOKENS: ${MAX_INPUT_TOKENS}
MAX_TOTAL_TOKENS: ${MAX_TOTAL_TOKENS} MAX_TOTAL_TOKENS: ${MAX_TOTAL_TOKENS}
LLM_ENDPOINT: ${LLM_ENDPOINT} LLM_ENDPOINT: ${LLM_ENDPOINT}

View File

@@ -13,7 +13,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0 HF_HUB_ENABLE_HF_TRANSFER: 0
HABANA_VISIBLE_DEVICES: all HABANA_VISIBLE_DEVICES: all
@@ -48,7 +48,7 @@ services:
no_proxy: ${no_proxy} no_proxy: ${no_proxy}
http_proxy: ${http_proxy} http_proxy: ${http_proxy}
https_proxy: ${https_proxy} https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
MAX_INPUT_TOKENS: ${MAX_INPUT_TOKENS} MAX_INPUT_TOKENS: ${MAX_INPUT_TOKENS}
MAX_TOTAL_TOKENS: ${MAX_TOTAL_TOKENS} MAX_TOTAL_TOKENS: ${MAX_TOTAL_TOKENS}
LLM_ENDPOINT: ${LLM_ENDPOINT} LLM_ENDPOINT: ${LLM_ENDPOINT}

View File

@@ -10,7 +10,7 @@ export host_ip=$(hostname -I | awk '{print $1}') # Example: host_ip="192.168.1.1
export no_proxy="${no_proxy},${host_ip}" # Example: no_proxy="localhost, 127.0.0.1, 192.168.1.1" export no_proxy="${no_proxy},${host_ip}" # Example: no_proxy="localhost, 127.0.0.1, 192.168.1.1"
export http_proxy=$http_proxy export http_proxy=$http_proxy
export https_proxy=$https_proxy export https_proxy=$https_proxy
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
export LLM_ENDPOINT_PORT=8008 export LLM_ENDPOINT_PORT=8008
export LLM_MODEL_ID="Intel/neural-chat-7b-v3-3" export LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"

Some files were not shown because too many files have changed in this diff Show More