Compare commits
1 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
3e29a76fc4 |
2
.github/workflows/_gmc-e2e.yml
vendored
2
.github/workflows/_gmc-e2e.yml
vendored
@@ -55,7 +55,7 @@ jobs:
|
|||||||
- name: Run tests
|
- name: Run tests
|
||||||
id: run-test
|
id: run-test
|
||||||
env:
|
env:
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${{ secrets.HUGGINGFACEHUB_API_TOKEN }}
|
HUGGINGFACEHUB_API_TOKEN: ${{ secrets.HF_TOKEN }}
|
||||||
GOOGLE_CSE_ID: ${{ secrets.GOOGLE_CSE_ID }}
|
GOOGLE_CSE_ID: ${{ secrets.GOOGLE_CSE_ID }}
|
||||||
GOOGLE_API_KEY: ${{ secrets.GOOGLE_API_KEY }}
|
GOOGLE_API_KEY: ${{ secrets.GOOGLE_API_KEY }}
|
||||||
run: |
|
run: |
|
||||||
|
|||||||
4
.github/workflows/_helm-e2e.yml
vendored
4
.github/workflows/_helm-e2e.yml
vendored
@@ -165,8 +165,8 @@ jobs:
|
|||||||
env:
|
env:
|
||||||
GOOGLE_CSE_ID: ${{ secrets.GOOGLE_CSE_ID }}
|
GOOGLE_CSE_ID: ${{ secrets.GOOGLE_CSE_ID }}
|
||||||
GOOGLE_API_KEY: ${{ secrets.GOOGLE_API_KEY }}
|
GOOGLE_API_KEY: ${{ secrets.GOOGLE_API_KEY }}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${{ secrets.HUGGINGFACEHUB_API_TOKEN }}
|
HUGGINGFACEHUB_API_TOKEN: ${{ secrets.HF_TOKEN }}
|
||||||
HFTOKEN: ${{ secrets.HUGGINGFACEHUB_API_TOKEN }}
|
HFTOKEN: ${{ secrets.HF_TOKEN }}
|
||||||
value_file: ${{ matrix.value_file }}
|
value_file: ${{ matrix.value_file }}
|
||||||
run: |
|
run: |
|
||||||
set -xe
|
set -xe
|
||||||
|
|||||||
4
.github/workflows/_run-docker-compose.yml
vendored
4
.github/workflows/_run-docker-compose.yml
vendored
@@ -160,8 +160,8 @@ jobs:
|
|||||||
- name: Run test
|
- name: Run test
|
||||||
shell: bash
|
shell: bash
|
||||||
env:
|
env:
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${{ secrets.HUGGINGFACEHUB_API_TOKEN }}
|
HUGGINGFACEHUB_API_TOKEN: ${{ secrets.HF_TOKEN }}
|
||||||
HF_TOKEN: ${{ secrets.HUGGINGFACEHUB_API_TOKEN }}
|
HF_TOKEN: ${{ secrets.HF_TOKEN }}
|
||||||
GOOGLE_CSE_ID: ${{ secrets.GOOGLE_CSE_ID }}
|
GOOGLE_CSE_ID: ${{ secrets.GOOGLE_CSE_ID }}
|
||||||
GOOGLE_API_KEY: ${{ secrets.GOOGLE_API_KEY }}
|
GOOGLE_API_KEY: ${{ secrets.GOOGLE_API_KEY }}
|
||||||
PINECONE_KEY: ${{ secrets.PINECONE_KEY }}
|
PINECONE_KEY: ${{ secrets.PINECONE_KEY }}
|
||||||
|
|||||||
@@ -13,8 +13,8 @@ services:
|
|||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
TGI_LLM_ENDPOINT: "http://${ip_address}:${TGI_SERVICE_PORT}"
|
TGI_LLM_ENDPOINT: "http://${ip_address}:${TGI_SERVICE_PORT}"
|
||||||
HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
shm_size: 32g
|
shm_size: 32g
|
||||||
devices:
|
devices:
|
||||||
- /dev/kfd:/dev/kfd
|
- /dev/kfd:/dev/kfd
|
||||||
@@ -42,7 +42,7 @@ services:
|
|||||||
with_memory: false
|
with_memory: false
|
||||||
recursion_limit: ${recursion_limit_worker}
|
recursion_limit: ${recursion_limit_worker}
|
||||||
llm_engine: tgi
|
llm_engine: tgi
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
llm_endpoint_url: ${LLM_ENDPOINT_URL}
|
llm_endpoint_url: ${LLM_ENDPOINT_URL}
|
||||||
model: ${LLM_MODEL_ID}
|
model: ${LLM_MODEL_ID}
|
||||||
temperature: ${temperature}
|
temperature: ${temperature}
|
||||||
@@ -76,7 +76,7 @@ services:
|
|||||||
use_hints: false
|
use_hints: false
|
||||||
recursion_limit: ${recursion_limit_worker}
|
recursion_limit: ${recursion_limit_worker}
|
||||||
llm_engine: vllm
|
llm_engine: vllm
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
llm_endpoint_url: ${LLM_ENDPOINT_URL}
|
llm_endpoint_url: ${LLM_ENDPOINT_URL}
|
||||||
model: ${LLM_MODEL_ID}
|
model: ${LLM_MODEL_ID}
|
||||||
temperature: ${temperature}
|
temperature: ${temperature}
|
||||||
@@ -104,7 +104,7 @@ services:
|
|||||||
with_memory: true
|
with_memory: true
|
||||||
recursion_limit: ${recursion_limit_supervisor}
|
recursion_limit: ${recursion_limit_supervisor}
|
||||||
llm_engine: tgi
|
llm_engine: tgi
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
llm_endpoint_url: ${LLM_ENDPOINT_URL}
|
llm_endpoint_url: ${LLM_ENDPOINT_URL}
|
||||||
model: ${LLM_MODEL_ID}
|
model: ${LLM_MODEL_ID}
|
||||||
temperature: ${temperature}
|
temperature: ${temperature}
|
||||||
|
|||||||
@@ -10,8 +10,8 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
WILM_USE_TRITON_FLASH_ATTENTION: 0
|
WILM_USE_TRITON_FLASH_ATTENTION: 0
|
||||||
@@ -46,7 +46,7 @@ services:
|
|||||||
with_memory: false
|
with_memory: false
|
||||||
recursion_limit: ${recursion_limit_worker}
|
recursion_limit: ${recursion_limit_worker}
|
||||||
llm_engine: vllm
|
llm_engine: vllm
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
llm_endpoint_url: ${LLM_ENDPOINT_URL}
|
llm_endpoint_url: ${LLM_ENDPOINT_URL}
|
||||||
model: ${LLM_MODEL_ID}
|
model: ${LLM_MODEL_ID}
|
||||||
temperature: ${temperature}
|
temperature: ${temperature}
|
||||||
@@ -80,7 +80,7 @@ services:
|
|||||||
use_hints: false
|
use_hints: false
|
||||||
recursion_limit: ${recursion_limit_worker}
|
recursion_limit: ${recursion_limit_worker}
|
||||||
llm_engine: vllm
|
llm_engine: vllm
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
llm_endpoint_url: ${LLM_ENDPOINT_URL}
|
llm_endpoint_url: ${LLM_ENDPOINT_URL}
|
||||||
model: ${LLM_MODEL_ID}
|
model: ${LLM_MODEL_ID}
|
||||||
temperature: ${temperature}
|
temperature: ${temperature}
|
||||||
@@ -108,7 +108,7 @@ services:
|
|||||||
with_memory: true
|
with_memory: true
|
||||||
recursion_limit: ${recursion_limit_supervisor}
|
recursion_limit: ${recursion_limit_supervisor}
|
||||||
llm_engine: vllm
|
llm_engine: vllm
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
llm_endpoint_url: ${LLM_ENDPOINT_URL}
|
llm_endpoint_url: ${LLM_ENDPOINT_URL}
|
||||||
model: ${LLM_MODEL_ID}
|
model: ${LLM_MODEL_ID}
|
||||||
temperature: ${temperature}
|
temperature: ${temperature}
|
||||||
|
|||||||
@@ -19,7 +19,7 @@ export CRAG_SERVER_PORT="18114"
|
|||||||
|
|
||||||
export WORKPATH=$(dirname "$PWD")
|
export WORKPATH=$(dirname "$PWD")
|
||||||
export WORKDIR=${WORKPATH}/../../../
|
export WORKDIR=${WORKPATH}/../../../
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
|
export LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
|
||||||
export HF_CACHE_DIR="./data"
|
export HF_CACHE_DIR="./data"
|
||||||
export MODEL_CACHE="./data"
|
export MODEL_CACHE="./data"
|
||||||
@@ -39,7 +39,7 @@ export CRAG_SERVER=http://${ip_address}:${CRAG_SERVER_PORT}
|
|||||||
export WORKER_AGENT_URL="http://${ip_address}:${WORKER_RAG_AGENT_PORT}/v1/chat/completions"
|
export WORKER_AGENT_URL="http://${ip_address}:${WORKER_RAG_AGENT_PORT}/v1/chat/completions"
|
||||||
export SQL_AGENT_URL="http://${ip_address}:${WORKER_SQL_AGENT_PORT}/v1/chat/completions"
|
export SQL_AGENT_URL="http://${ip_address}:${WORKER_SQL_AGENT_PORT}/v1/chat/completions"
|
||||||
export HF_CACHE_DIR=${HF_CACHE_DIR}
|
export HF_CACHE_DIR=${HF_CACHE_DIR}
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export no_proxy=${no_proxy}
|
export no_proxy=${no_proxy}
|
||||||
export http_proxy=${http_proxy}
|
export http_proxy=${http_proxy}
|
||||||
export https_proxy=${https_proxy}
|
export https_proxy=${https_proxy}
|
||||||
|
|||||||
@@ -19,7 +19,7 @@ export CRAG_SERVER_PORT="18114"
|
|||||||
|
|
||||||
export WORKPATH=$(dirname "$PWD")
|
export WORKPATH=$(dirname "$PWD")
|
||||||
export WORKDIR=${WORKPATH}/../../../
|
export WORKDIR=${WORKPATH}/../../../
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export VLLM_LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
|
export VLLM_LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
|
||||||
export HF_CACHE_DIR="./data"
|
export HF_CACHE_DIR="./data"
|
||||||
export MODEL_CACHE="./data"
|
export MODEL_CACHE="./data"
|
||||||
@@ -40,7 +40,7 @@ export CRAG_SERVER=http://${ip_address}:${CRAG_SERVER_PORT}
|
|||||||
export WORKER_AGENT_URL="http://${ip_address}:${WORKER_RAG_AGENT_PORT}/v1/chat/completions"
|
export WORKER_AGENT_URL="http://${ip_address}:${WORKER_RAG_AGENT_PORT}/v1/chat/completions"
|
||||||
export SQL_AGENT_URL="http://${ip_address}:${WORKER_SQL_AGENT_PORT}/v1/chat/completions"
|
export SQL_AGENT_URL="http://${ip_address}:${WORKER_SQL_AGENT_PORT}/v1/chat/completions"
|
||||||
export HF_CACHE_DIR=${HF_CACHE_DIR}
|
export HF_CACHE_DIR=${HF_CACHE_DIR}
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export no_proxy=${no_proxy}
|
export no_proxy=${no_proxy}
|
||||||
export http_proxy=${http_proxy}
|
export http_proxy=${http_proxy}
|
||||||
export https_proxy=${https_proxy}
|
export https_proxy=${https_proxy}
|
||||||
|
|||||||
@@ -20,8 +20,8 @@ export CRAG_SERVER_PORT="18114"
|
|||||||
|
|
||||||
export WORKPATH=$(dirname "$PWD")
|
export WORKPATH=$(dirname "$PWD")
|
||||||
export WORKDIR=${WORKPATH}/../../../
|
export WORKDIR=${WORKPATH}/../../../
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export VLLM_LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
|
export VLLM_LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
|
||||||
export HF_CACHE_DIR="./data"
|
export HF_CACHE_DIR="./data"
|
||||||
export MODEL_CACHE="./data"
|
export MODEL_CACHE="./data"
|
||||||
@@ -42,7 +42,7 @@ export CRAG_SERVER=http://${ip_address}:${CRAG_SERVER_PORT}
|
|||||||
export WORKER_AGENT_URL="http://${ip_address}:${WORKER_RAG_AGENT_PORT}/v1/chat/completions"
|
export WORKER_AGENT_URL="http://${ip_address}:${WORKER_RAG_AGENT_PORT}/v1/chat/completions"
|
||||||
export SQL_AGENT_URL="http://${ip_address}:${WORKER_SQL_AGENT_PORT}/v1/chat/completions"
|
export SQL_AGENT_URL="http://${ip_address}:${WORKER_SQL_AGENT_PORT}/v1/chat/completions"
|
||||||
export HF_CACHE_DIR=${HF_CACHE_DIR}
|
export HF_CACHE_DIR=${HF_CACHE_DIR}
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export no_proxy=${no_proxy}
|
export no_proxy=${no_proxy}
|
||||||
export http_proxy=${http_proxy}
|
export http_proxy=${http_proxy}
|
||||||
export https_proxy=${https_proxy}
|
export https_proxy=${https_proxy}
|
||||||
|
|||||||
@@ -33,7 +33,7 @@ fi
|
|||||||
# retriever
|
# retriever
|
||||||
export host_ip=$(hostname -I | awk '{print $1}')
|
export host_ip=$(hostname -I | awk '{print $1}')
|
||||||
export HF_CACHE_DIR=${HF_CACHE_DIR}
|
export HF_CACHE_DIR=${HF_CACHE_DIR}
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export no_proxy=${no_proxy}
|
export no_proxy=${no_proxy}
|
||||||
export http_proxy=${http_proxy}
|
export http_proxy=${http_proxy}
|
||||||
export https_proxy=${https_proxy}
|
export https_proxy=${https_proxy}
|
||||||
|
|||||||
@@ -16,7 +16,7 @@ services:
|
|||||||
with_memory: false
|
with_memory: false
|
||||||
recursion_limit: ${recursion_limit_worker}
|
recursion_limit: ${recursion_limit_worker}
|
||||||
llm_engine: vllm
|
llm_engine: vllm
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
llm_endpoint_url: ${LLM_ENDPOINT_URL}
|
llm_endpoint_url: ${LLM_ENDPOINT_URL}
|
||||||
model: ${LLM_MODEL_ID}
|
model: ${LLM_MODEL_ID}
|
||||||
temperature: ${temperature}
|
temperature: ${temperature}
|
||||||
@@ -50,7 +50,7 @@ services:
|
|||||||
use_hints: false
|
use_hints: false
|
||||||
recursion_limit: ${recursion_limit_worker}
|
recursion_limit: ${recursion_limit_worker}
|
||||||
llm_engine: vllm
|
llm_engine: vllm
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
llm_endpoint_url: ${LLM_ENDPOINT_URL}
|
llm_endpoint_url: ${LLM_ENDPOINT_URL}
|
||||||
model: ${LLM_MODEL_ID}
|
model: ${LLM_MODEL_ID}
|
||||||
temperature: ${temperature}
|
temperature: ${temperature}
|
||||||
@@ -79,7 +79,7 @@ services:
|
|||||||
with_memory: true
|
with_memory: true
|
||||||
recursion_limit: ${recursion_limit_supervisor}
|
recursion_limit: ${recursion_limit_supervisor}
|
||||||
llm_engine: vllm
|
llm_engine: vllm
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
llm_endpoint_url: ${LLM_ENDPOINT_URL}
|
llm_endpoint_url: ${LLM_ENDPOINT_URL}
|
||||||
model: ${LLM_MODEL_ID}
|
model: ${LLM_MODEL_ID}
|
||||||
temperature: ${temperature}
|
temperature: ${temperature}
|
||||||
@@ -122,7 +122,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
HABANA_VISIBLE_DEVICES: all
|
HABANA_VISIBLE_DEVICES: all
|
||||||
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
|
|||||||
@@ -16,8 +16,8 @@ export ip_address=$(hostname -I | awk '{print $1}')
|
|||||||
# LLM related environment variables
|
# LLM related environment variables
|
||||||
export HF_CACHE_DIR=${HF_CACHE_DIR}
|
export HF_CACHE_DIR=${HF_CACHE_DIR}
|
||||||
ls $HF_CACHE_DIR
|
ls $HF_CACHE_DIR
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export HF_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HF_TOKEN=${HF_TOKEN}
|
||||||
export LLM_MODEL_ID="meta-llama/Llama-3.3-70B-Instruct"
|
export LLM_MODEL_ID="meta-llama/Llama-3.3-70B-Instruct"
|
||||||
export NUM_SHARDS=4
|
export NUM_SHARDS=4
|
||||||
export LLM_ENDPOINT_URL="http://${ip_address}:8086"
|
export LLM_ENDPOINT_URL="http://${ip_address}:8086"
|
||||||
|
|||||||
@@ -13,7 +13,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
HABANA_VISIBLE_DEVICES: all
|
HABANA_VISIBLE_DEVICES: all
|
||||||
|
|||||||
@@ -3,7 +3,7 @@
|
|||||||
|
|
||||||
host_ip=$(hostname -I | awk '{print $1}')
|
host_ip=$(hostname -I | awk '{print $1}')
|
||||||
export HF_CACHE_DIR=${HF_CACHE_DIR}
|
export HF_CACHE_DIR=${HF_CACHE_DIR}
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export no_proxy=${no_proxy}
|
export no_proxy=${no_proxy}
|
||||||
export http_proxy=${http_proxy}
|
export http_proxy=${http_proxy}
|
||||||
export https_proxy=${https_proxy}
|
export https_proxy=${https_proxy}
|
||||||
|
|||||||
@@ -20,7 +20,7 @@ function start_retrieval_tool() {
|
|||||||
cd $WORKPATH/../DocIndexRetriever/docker_compose/intel/cpu/xeon
|
cd $WORKPATH/../DocIndexRetriever/docker_compose/intel/cpu/xeon
|
||||||
host_ip=$(hostname -I | awk '{print $1}')
|
host_ip=$(hostname -I | awk '{print $1}')
|
||||||
export HF_CACHE_DIR=${HF_CACHE_DIR}
|
export HF_CACHE_DIR=${HF_CACHE_DIR}
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export no_proxy=${no_proxy}
|
export no_proxy=${no_proxy}
|
||||||
export http_proxy=${http_proxy}
|
export http_proxy=${http_proxy}
|
||||||
export https_proxy=${https_proxy}
|
export https_proxy=${https_proxy}
|
||||||
|
|||||||
@@ -11,8 +11,8 @@ export ip_address=$(hostname -I | awk '{print $1}')
|
|||||||
export host_ip=$ip_address
|
export host_ip=$ip_address
|
||||||
echo "ip_address=${ip_address}"
|
echo "ip_address=${ip_address}"
|
||||||
export TOOLSET_PATH=$WORKPATH/tools/
|
export TOOLSET_PATH=$WORKPATH/tools/
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
HF_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN=${HF_TOKEN}
|
||||||
model="meta-llama/Llama-3.3-70B-Instruct" #"meta-llama/Meta-Llama-3.1-70B-Instruct"
|
model="meta-llama/Llama-3.3-70B-Instruct" #"meta-llama/Meta-Llama-3.1-70B-Instruct"
|
||||||
|
|
||||||
export HF_CACHE_DIR=${model_cache:-"/data2/huggingface"}
|
export HF_CACHE_DIR=${model_cache:-"/data2/huggingface"}
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ WORKPATH=$(dirname "$PWD")
|
|||||||
export WORKDIR=$WORKPATH/../../
|
export WORKDIR=$WORKPATH/../../
|
||||||
echo "WORKDIR=${WORKDIR}"
|
echo "WORKDIR=${WORKDIR}"
|
||||||
export ip_address=$(hostname -I | awk '{print $1}')
|
export ip_address=$(hostname -I | awk '{print $1}')
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export TOOLSET_PATH=$WORKDIR/GenAIExamples/AgentQnA/tools/
|
export TOOLSET_PATH=$WORKDIR/GenAIExamples/AgentQnA/tools/
|
||||||
export no_proxy="$no_proxy,rag-agent-endpoint,sql-agent-endpoint,react-agent-endpoint,agent-ui,vllm-gaudi-server,jaeger,grafana,prometheus,127.0.0.1,localhost,0.0.0.0,$ip_address"
|
export no_proxy="$no_proxy,rag-agent-endpoint,sql-agent-endpoint,react-agent-endpoint,agent-ui,vllm-gaudi-server,jaeger,grafana,prometheus,127.0.0.1,localhost,0.0.0.0,$ip_address"
|
||||||
IMAGE_REPO=${IMAGE_REPO:-"opea"}
|
IMAGE_REPO=${IMAGE_REPO:-"opea"}
|
||||||
|
|||||||
@@ -9,7 +9,7 @@ ls $WORKPATH
|
|||||||
export WORKDIR=$WORKPATH/../../
|
export WORKDIR=$WORKPATH/../../
|
||||||
echo "WORKDIR=${WORKDIR}"
|
echo "WORKDIR=${WORKDIR}"
|
||||||
export ip_address=$(hostname -I | awk '{print $1}')
|
export ip_address=$(hostname -I | awk '{print $1}')
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export TOOLSET_PATH=$WORKPATH/tools/
|
export TOOLSET_PATH=$WORKPATH/tools/
|
||||||
IMAGE_REPO=${IMAGE_REPO:-"opea"}
|
IMAGE_REPO=${IMAGE_REPO:-"opea"}
|
||||||
IMAGE_TAG=${IMAGE_TAG:-"latest"}
|
IMAGE_TAG=${IMAGE_TAG:-"latest"}
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ WORKPATH=$(dirname "$PWD")
|
|||||||
export WORKDIR=${WORKPATH}/../../
|
export WORKDIR=${WORKPATH}/../../
|
||||||
echo "WORKDIR=${WORKDIR}"
|
echo "WORKDIR=${WORKDIR}"
|
||||||
export ip_address=$(hostname -I | awk '{print $1}')
|
export ip_address=$(hostname -I | awk '{print $1}')
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export TOOLSET_PATH=$WORKPATH/tools/
|
export TOOLSET_PATH=$WORKPATH/tools/
|
||||||
IMAGE_REPO=${IMAGE_REPO:-"opea"}
|
IMAGE_REPO=${IMAGE_REPO:-"opea"}
|
||||||
IMAGE_TAG=${IMAGE_TAG:-"latest"}
|
IMAGE_TAG=${IMAGE_TAG:-"latest"}
|
||||||
|
|||||||
@@ -40,7 +40,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
command: --model-id ${LLM_MODEL_ID}
|
command: --model-id ${LLM_MODEL_ID}
|
||||||
|
|||||||
@@ -35,8 +35,8 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
WILM_USE_TRITON_FLASH_ATTENTION: 0
|
WILM_USE_TRITON_FLASH_ATTENTION: 0
|
||||||
|
|||||||
@@ -7,7 +7,7 @@
|
|||||||
# export host_ip=<your External Public IP> # export host_ip=$(hostname -I | awk '{print $1}')
|
# export host_ip=<your External Public IP> # export host_ip=$(hostname -I | awk '{print $1}')
|
||||||
|
|
||||||
export host_ip=${ip_address}
|
export host_ip=${ip_address}
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
# <token>
|
# <token>
|
||||||
|
|
||||||
export LLM_MODEL_ID=Intel/neural-chat-7b-v3-3
|
export LLM_MODEL_ID=Intel/neural-chat-7b-v3-3
|
||||||
|
|||||||
@@ -8,7 +8,7 @@
|
|||||||
|
|
||||||
export host_ip=${ip_address}
|
export host_ip=${ip_address}
|
||||||
export external_host_ip=${ip_address}
|
export external_host_ip=${ip_address}
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export HF_CACHE_DIR="./data"
|
export HF_CACHE_DIR="./data"
|
||||||
export LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
|
export LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
|
||||||
export VLLM_SERVICE_PORT="8081"
|
export VLLM_SERVICE_PORT="8081"
|
||||||
|
|||||||
@@ -36,7 +36,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
VLLM_TORCH_PROFILER_DIR: "/mnt"
|
VLLM_TORCH_PROFILER_DIR: "/mnt"
|
||||||
LLM_SERVER_PORT: ${LLM_SERVER_PORT}
|
LLM_SERVER_PORT: ${LLM_SERVER_PORT}
|
||||||
|
|||||||
@@ -40,7 +40,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
VLLM_TORCH_PROFILER_DIR: "/mnt"
|
VLLM_TORCH_PROFILER_DIR: "/mnt"
|
||||||
LLM_SERVER_PORT: ${LLM_SERVER_PORT}
|
LLM_SERVER_PORT: ${LLM_SERVER_PORT}
|
||||||
|
|||||||
@@ -36,7 +36,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
LLM_SERVER_PORT: ${LLM_SERVER_PORT}
|
LLM_SERVER_PORT: ${LLM_SERVER_PORT}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://$host_ip:${LLM_SERVER_PORT}/health || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://$host_ip:${LLM_SERVER_PORT}/health || exit 1"]
|
||||||
|
|||||||
@@ -5,7 +5,7 @@
|
|||||||
|
|
||||||
# export host_ip=<your External Public IP>
|
# export host_ip=<your External Public IP>
|
||||||
export host_ip=$(hostname -I | awk '{print $1}')
|
export host_ip=$(hostname -I | awk '{print $1}')
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
# <token>
|
# <token>
|
||||||
|
|
||||||
export LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
|
export LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
|
||||||
|
|||||||
@@ -45,7 +45,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
HABANA_VISIBLE_DEVICES: all
|
HABANA_VISIBLE_DEVICES: all
|
||||||
|
|||||||
@@ -45,7 +45,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
HABANA_VISIBLE_DEVICES: all
|
HABANA_VISIBLE_DEVICES: all
|
||||||
|
|||||||
@@ -5,7 +5,7 @@
|
|||||||
|
|
||||||
# export host_ip=<your External Public IP>
|
# export host_ip=<your External Public IP>
|
||||||
export host_ip=$(hostname -I | awk '{print $1}')
|
export host_ip=$(hostname -I | awk '{print $1}')
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
# <token>
|
# <token>
|
||||||
|
|
||||||
export LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
|
export LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
|
||||||
|
|||||||
@@ -68,7 +68,7 @@ Then run the command `docker images`, you will have following images ready:
|
|||||||
Before starting the services with `docker compose`, you have to recheck the following environment variables.
|
Before starting the services with `docker compose`, you have to recheck the following environment variables.
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export host_ip=$(hostname -I | awk '{print $1}')
|
export host_ip=$(hostname -I | awk '{print $1}')
|
||||||
|
|
||||||
export TGI_SERVICE_PORT=3006
|
export TGI_SERVICE_PORT=3006
|
||||||
|
|||||||
@@ -52,8 +52,8 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
shm_size: 1g
|
shm_size: 1g
|
||||||
devices:
|
devices:
|
||||||
- /dev/kfd:/dev/kfd
|
- /dev/kfd:/dev/kfd
|
||||||
|
|||||||
@@ -3,7 +3,7 @@
|
|||||||
# Copyright (C) 2024 Intel Corporation
|
# Copyright (C) 2024 Intel Corporation
|
||||||
# SPDX-License-Identifier: Apache-2.0
|
# SPDX-License-Identifier: Apache-2.0
|
||||||
|
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export OPENAI_API_KEY=${OPENAI_API_KEY}
|
export OPENAI_API_KEY=${OPENAI_API_KEY}
|
||||||
export host_ip=$(hostname -I | awk '{print $1}')
|
export host_ip=$(hostname -I | awk '{print $1}')
|
||||||
|
|
||||||
|
|||||||
@@ -37,7 +37,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://${host_ip}:3006/health || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://${host_ip}:3006/health || exit 1"]
|
||||||
interval: 10s
|
interval: 10s
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ pushd "../../../../../" > /dev/null
|
|||||||
source .set_env.sh
|
source .set_env.sh
|
||||||
popd > /dev/null
|
popd > /dev/null
|
||||||
|
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export host_ip=$(hostname -I | awk '{print $1}')
|
export host_ip=$(hostname -I | awk '{print $1}')
|
||||||
export LLM_MODEL_ID=Intel/neural-chat-7b-v3-3
|
export LLM_MODEL_ID=Intel/neural-chat-7b-v3-3
|
||||||
export WAV2LIP_ENDPOINT=http://$host_ip:7860
|
export WAV2LIP_ENDPOINT=http://$host_ip:7860
|
||||||
|
|||||||
@@ -48,7 +48,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
HABANA_VISIBLE_DEVICES: all
|
HABANA_VISIBLE_DEVICES: all
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ pushd "../../../../../" > /dev/null
|
|||||||
source .set_env.sh
|
source .set_env.sh
|
||||||
popd > /dev/null
|
popd > /dev/null
|
||||||
|
|
||||||
export HUGGINGFACEHUB_API_TOKEN=$HUGGINGFACEHUB_API_TOKEN
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export host_ip=$(hostname -I | awk '{print $1}')
|
export host_ip=$(hostname -I | awk '{print $1}')
|
||||||
|
|
||||||
export LLM_MODEL_ID=Intel/neural-chat-7b-v3-3
|
export LLM_MODEL_ID=Intel/neural-chat-7b-v3-3
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ export HOST_IP=${ip_address}
|
|||||||
export HOST_IP_EXTERNAL=${ip_address}
|
export HOST_IP_EXTERNAL=${ip_address}
|
||||||
|
|
||||||
export CHATQNA_EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
|
export CHATQNA_EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
|
||||||
export CHATQNA_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export CHATQNA_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export CHATQNA_LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
|
export CHATQNA_LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
|
||||||
export CHATQNA_RERANK_MODEL_ID="BAAI/bge-reranker-base"
|
export CHATQNA_RERANK_MODEL_ID="BAAI/bge-reranker-base"
|
||||||
|
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ export HOST_IP=${ip_address}
|
|||||||
export HOST_IP_EXTERNAL=${ip_address}
|
export HOST_IP_EXTERNAL=${ip_address}
|
||||||
|
|
||||||
export CHATQNA_EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
|
export CHATQNA_EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
|
||||||
export CHATQNA_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export CHATQNA_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export CHATQNA_LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
|
export CHATQNA_LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
|
||||||
export CHATQNA_RERANK_MODEL_ID="BAAI/bge-reranker-base"
|
export CHATQNA_RERANK_MODEL_ID="BAAI/bge-reranker-base"
|
||||||
|
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ export HOST_IP=${ip_address}
|
|||||||
export HOST_IP_EXTERNAL=${ip_address}
|
export HOST_IP_EXTERNAL=${ip_address}
|
||||||
|
|
||||||
export CHATQNA_EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
|
export CHATQNA_EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
|
||||||
export CHATQNA_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export CHATQNA_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export CHATQNA_LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
|
export CHATQNA_LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
|
||||||
export CHATQNA_RERANK_MODEL_ID="BAAI/bge-reranker-base"
|
export CHATQNA_RERANK_MODEL_ID="BAAI/bge-reranker-base"
|
||||||
|
|
||||||
|
|||||||
@@ -6,7 +6,7 @@ export HOST_IP=${ip_address}
|
|||||||
export HOST_IP_EXTERNAL=${ip_address}
|
export HOST_IP_EXTERNAL=${ip_address}
|
||||||
|
|
||||||
export CHATQNA_EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
|
export CHATQNA_EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
|
||||||
export CHATQNA_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export CHATQNA_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export CHATQNA_LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
|
export CHATQNA_LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
|
||||||
export CHATQNA_RERANK_MODEL_ID="BAAI/bge-reranker-base"
|
export CHATQNA_RERANK_MODEL_ID="BAAI/bge-reranker-base"
|
||||||
|
|
||||||
|
|||||||
@@ -183,7 +183,7 @@ export https_proxy=${your_http_proxy}
|
|||||||
export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
|
export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
|
||||||
export RERANK_MODEL_ID="BAAI/bge-reranker-base"
|
export RERANK_MODEL_ID="BAAI/bge-reranker-base"
|
||||||
export INDEX_NAME="rag-redis"
|
export INDEX_NAME="rag-redis"
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export OLLAMA_HOST=${host_ip}
|
export OLLAMA_HOST=${host_ip}
|
||||||
export OLLAMA_MODEL="llama3.2"
|
export OLLAMA_MODEL="llama3.2"
|
||||||
```
|
```
|
||||||
@@ -194,7 +194,7 @@ export OLLAMA_MODEL="llama3.2"
|
|||||||
set EMBEDDING_MODEL_ID=BAAI/bge-base-en-v1.5
|
set EMBEDDING_MODEL_ID=BAAI/bge-base-en-v1.5
|
||||||
set RERANK_MODEL_ID=BAAI/bge-reranker-base
|
set RERANK_MODEL_ID=BAAI/bge-reranker-base
|
||||||
set INDEX_NAME=rag-redis
|
set INDEX_NAME=rag-redis
|
||||||
set HUGGINGFACEHUB_API_TOKEN=%HUGGINGFACEHUB_API_TOKEN%
|
set HUGGINGFACEHUB_API_TOKEN=%HF_TOKEN%
|
||||||
set OLLAMA_HOST=host.docker.internal
|
set OLLAMA_HOST=host.docker.internal
|
||||||
set OLLAMA_MODEL="llama3.2"
|
set OLLAMA_MODEL="llama3.2"
|
||||||
```
|
```
|
||||||
|
|||||||
@@ -24,7 +24,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_ENDPOINT: http://tei-embedding-service:80
|
TEI_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
interval: 10s
|
interval: 10s
|
||||||
@@ -60,7 +60,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
@@ -76,7 +76,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
||||||
|
|||||||
@@ -9,7 +9,7 @@ popd > /dev/null
|
|||||||
|
|
||||||
export host_ip=$(hostname -I | awk '{print $1}')
|
export host_ip=$(hostname -I | awk '{print $1}')
|
||||||
|
|
||||||
if [ -z "${HUGGINGFACEHUB_API_TOKEN}" ]; then
|
if [ -z "${HF_TOKEN}" ]; then
|
||||||
echo "Error: HUGGINGFACEHUB_API_TOKEN is not set. Please set HUGGINGFACEHUB_API_TOKEN."
|
echo "Error: HUGGINGFACEHUB_API_TOKEN is not set. Please set HUGGINGFACEHUB_API_TOKEN."
|
||||||
fi
|
fi
|
||||||
|
|
||||||
@@ -17,7 +17,7 @@ if [ -z "${host_ip}" ]; then
|
|||||||
echo "Error: host_ip is not set. Please set host_ip first."
|
echo "Error: host_ip is not set. Please set host_ip first."
|
||||||
fi
|
fi
|
||||||
|
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
|
export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
|
||||||
export RERANK_MODEL_ID="BAAI/bge-reranker-base"
|
export RERANK_MODEL_ID="BAAI/bge-reranker-base"
|
||||||
export INDEX_NAME="rag-redis"
|
export INDEX_NAME="rag-redis"
|
||||||
|
|||||||
@@ -31,7 +31,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_ENDPOINT: http://tei-embedding-service:80
|
TEI_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
interval: 10s
|
interval: 10s
|
||||||
@@ -67,7 +67,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
@@ -83,7 +83,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
||||||
@@ -99,7 +99,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
VLLM_TORCH_PROFILER_DIR: "/mnt"
|
VLLM_TORCH_PROFILER_DIR: "/mnt"
|
||||||
VLLM_CPU_KVCACHE_SPACE: 40
|
VLLM_CPU_KVCACHE_SPACE: 40
|
||||||
|
|||||||
@@ -24,7 +24,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_ENDPOINT: http://tei-embedding-service:80
|
TEI_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
interval: 10s
|
interval: 10s
|
||||||
@@ -60,7 +60,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
@@ -76,7 +76,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
||||||
|
|||||||
@@ -24,7 +24,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_ENDPOINT: http://tei-embedding-service:80
|
TEI_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
interval: 10s
|
interval: 10s
|
||||||
@@ -60,7 +60,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
@@ -76,7 +76,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
||||||
|
|||||||
@@ -35,7 +35,7 @@ services:
|
|||||||
DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_MARIADBVECTOR"
|
DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_MARIADBVECTOR"
|
||||||
MARIADB_CONNECTION_URL: mariadb+mariadbconnector://${MARIADB_USER}:${MARIADB_PASSWORD}@mariadb-server:3306/${MARIADB_DATABASE}
|
MARIADB_CONNECTION_URL: mariadb+mariadbconnector://${MARIADB_USER}:${MARIADB_PASSWORD}@mariadb-server:3306/${MARIADB_DATABASE}
|
||||||
TEI_ENDPOINT: http://tei-embedding-service:80
|
TEI_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
interval: 10s
|
interval: 10s
|
||||||
@@ -69,7 +69,7 @@ services:
|
|||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
MARIADB_CONNECTION_URL: mariadb+mariadbconnector://${MARIADB_USER}:${MARIADB_PASSWORD}@mariadb-server:3306/${MARIADB_DATABASE}
|
MARIADB_CONNECTION_URL: mariadb+mariadbconnector://${MARIADB_USER}:${MARIADB_PASSWORD}@mariadb-server:3306/${MARIADB_DATABASE}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_MARIADBVECTOR"
|
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_MARIADBVECTOR"
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
@@ -85,7 +85,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
||||||
@@ -101,7 +101,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
VLLM_TORCH_PROFILER_DIR: "/mnt"
|
VLLM_TORCH_PROFILER_DIR: "/mnt"
|
||||||
VLLM_CPU_KVCACHE_SPACE: 40
|
VLLM_CPU_KVCACHE_SPACE: 40
|
||||||
|
|||||||
@@ -75,7 +75,7 @@ services:
|
|||||||
MILVUS_HOST: ${host_ip}
|
MILVUS_HOST: ${host_ip}
|
||||||
MILVUS_PORT: 19530
|
MILVUS_PORT: 19530
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
EMBEDDING_MODEL_ID: ${EMBEDDING_MODEL_ID}
|
EMBEDDING_MODEL_ID: ${EMBEDDING_MODEL_ID}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
@@ -107,7 +107,7 @@ services:
|
|||||||
MILVUS_HOST: ${host_ip}
|
MILVUS_HOST: ${host_ip}
|
||||||
MILVUS_PORT: 19530
|
MILVUS_PORT: 19530
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_MILVUS"
|
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_MILVUS"
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
@@ -138,7 +138,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
||||||
@@ -155,7 +155,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
VLLM_TORCH_PROFILER_DIR: "/mnt"
|
VLLM_TORCH_PROFILER_DIR: "/mnt"
|
||||||
healthcheck:
|
healthcheck:
|
||||||
|
|||||||
@@ -20,7 +20,7 @@ services:
|
|||||||
PINECONE_INDEX_NAME: ${PINECONE_INDEX_NAME}
|
PINECONE_INDEX_NAME: ${PINECONE_INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY}
|
LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_PINECONE"
|
DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_PINECONE"
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
@@ -55,7 +55,7 @@ services:
|
|||||||
PINECONE_INDEX_NAME: ${PINECONE_INDEX_NAME}
|
PINECONE_INDEX_NAME: ${PINECONE_INDEX_NAME}
|
||||||
LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY}
|
LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY}
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_PINECONE"
|
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_PINECONE"
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
@@ -71,7 +71,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
||||||
@@ -87,7 +87,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
VLLM_TORCH_PROFILER_DIR: "/mnt"
|
VLLM_TORCH_PROFILER_DIR: "/mnt"
|
||||||
command: --model $LLM_MODEL_ID --host 0.0.0.0 --port 80
|
command: --model $LLM_MODEL_ID --host 0.0.0.0 --port 80
|
||||||
|
|||||||
@@ -24,7 +24,7 @@ services:
|
|||||||
QDRANT_PORT: 6333
|
QDRANT_PORT: 6333
|
||||||
QDRANT_INDEX_NAME: ${INDEX_NAME}
|
QDRANT_INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_QDRANT"
|
DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_QDRANT"
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
@@ -76,7 +76,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
||||||
@@ -92,7 +92,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
VLLM_TORCH_PROFILER_DIR: "/mnt"
|
VLLM_TORCH_PROFILER_DIR: "/mnt"
|
||||||
command: --model $LLM_MODEL_ID --host 0.0.0.0 --port 80
|
command: --model $LLM_MODEL_ID --host 0.0.0.0 --port 80
|
||||||
|
|||||||
@@ -31,7 +31,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_ENDPOINT: http://tei-embedding-service:80
|
TEI_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
tei-embedding-service:
|
tei-embedding-service:
|
||||||
image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5
|
image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5
|
||||||
container_name: tei-embedding-server
|
container_name: tei-embedding-server
|
||||||
@@ -61,7 +61,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
@@ -77,7 +77,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
||||||
|
|||||||
@@ -24,7 +24,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_ENDPOINT: http://tei-embedding-service:80
|
TEI_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
interval: 10s
|
interval: 10s
|
||||||
@@ -60,7 +60,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
@@ -76,7 +76,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
||||||
@@ -92,7 +92,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
command: --model-id ${LLM_MODEL_ID} --cuda-graphs 0
|
command: --model-id ${LLM_MODEL_ID} --cuda-graphs 0
|
||||||
|
|||||||
@@ -24,7 +24,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_ENDPOINT: http://tei-embedding-service:80
|
TEI_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
interval: 10s
|
interval: 10s
|
||||||
@@ -60,7 +60,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
@@ -76,7 +76,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
VLLM_TORCH_PROFILER_DIR: "/mnt"
|
VLLM_TORCH_PROFILER_DIR: "/mnt"
|
||||||
command: --model $LLM_MODEL_ID --host 0.0.0.0 --port 80
|
command: --model $LLM_MODEL_ID --host 0.0.0.0 --port 80
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ pushd "../../../../../" > /dev/null
|
|||||||
source .set_env.sh
|
source .set_env.sh
|
||||||
popd > /dev/null
|
popd > /dev/null
|
||||||
|
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export HF_TOKEN=${HF_TOKEN}
|
export HF_TOKEN=${HF_TOKEN}
|
||||||
export host_ip=${ip_address}
|
export host_ip=${ip_address}
|
||||||
export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
|
export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ pushd "../../../../../" > /dev/null
|
|||||||
source .set_env.sh
|
source .set_env.sh
|
||||||
popd > /dev/null
|
popd > /dev/null
|
||||||
|
|
||||||
if [ -z "${HUGGINGFACEHUB_API_TOKEN}" ]; then
|
if [ -z "${HF_TOKEN}" ]; then
|
||||||
echo "Error: HUGGINGFACEHUB_API_TOKEN is not set. Please set HUGGINGFACEHUB_API_TOKEN."
|
echo "Error: HUGGINGFACEHUB_API_TOKEN is not set. Please set HUGGINGFACEHUB_API_TOKEN."
|
||||||
fi
|
fi
|
||||||
|
|
||||||
@@ -15,7 +15,7 @@ export host_ip=$(hostname -I | awk '{print $1}')
|
|||||||
export MARIADB_DATABASE="vectordb"
|
export MARIADB_DATABASE="vectordb"
|
||||||
export MARIADB_USER="chatqna"
|
export MARIADB_USER="chatqna"
|
||||||
export MARIADB_PASSWORD="password"
|
export MARIADB_PASSWORD="password"
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
|
export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
|
||||||
export RERANK_MODEL_ID="BAAI/bge-reranker-base"
|
export RERANK_MODEL_ID="BAAI/bge-reranker-base"
|
||||||
export LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
|
export LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
|
||||||
|
|||||||
@@ -31,7 +31,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_ENDPOINT: http://tei-embedding-service:80
|
TEI_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
interval: 10s
|
interval: 10s
|
||||||
@@ -67,7 +67,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
tei-reranking-service:
|
tei-reranking-service:
|
||||||
image: ghcr.io/huggingface/tei-gaudi:1.5.0
|
image: ghcr.io/huggingface/tei-gaudi:1.5.0
|
||||||
@@ -101,7 +101,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
HABANA_VISIBLE_DEVICES: all
|
HABANA_VISIBLE_DEVICES: all
|
||||||
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
|
|||||||
@@ -24,7 +24,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_ENDPOINT: http://tei-embedding-service:80
|
TEI_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
@@ -61,7 +61,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
tei-reranking-service:
|
tei-reranking-service:
|
||||||
image: ghcr.io/huggingface/tei-gaudi:1.5.0
|
image: ghcr.io/huggingface/tei-gaudi:1.5.0
|
||||||
|
|||||||
@@ -24,7 +24,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_ENDPOINT: http://tei-embedding-service:80
|
TEI_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
@@ -61,7 +61,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
tei-reranking-service:
|
tei-reranking-service:
|
||||||
image: ghcr.io/huggingface/tei-gaudi:1.5.0
|
image: ghcr.io/huggingface/tei-gaudi:1.5.0
|
||||||
@@ -95,7 +95,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
HABANA_VISIBLE_DEVICES: all
|
HABANA_VISIBLE_DEVICES: all
|
||||||
|
|||||||
@@ -24,7 +24,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_ENDPOINT: http://tei-embedding-service:80
|
TEI_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
interval: 10s
|
interval: 10s
|
||||||
@@ -42,7 +42,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
HABANA_VISIBLE_DEVICES: all
|
HABANA_VISIBLE_DEVICES: all
|
||||||
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
||||||
GURADRAILS_MODEL_ID: ${GURADRAILS_MODEL_ID}
|
GURADRAILS_MODEL_ID: ${GURADRAILS_MODEL_ID}
|
||||||
@@ -73,7 +73,7 @@ services:
|
|||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
SAFETY_GUARD_MODEL_ID: ${GURADRAILS_MODEL_ID}
|
SAFETY_GUARD_MODEL_ID: ${GURADRAILS_MODEL_ID}
|
||||||
SAFETY_GUARD_ENDPOINT: http://vllm-guardrails-service:80
|
SAFETY_GUARD_ENDPOINT: http://vllm-guardrails-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
tei-embedding-service:
|
tei-embedding-service:
|
||||||
image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5
|
image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5
|
||||||
@@ -104,7 +104,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
@@ -140,7 +140,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
HABANA_VISIBLE_DEVICES: all
|
HABANA_VISIBLE_DEVICES: all
|
||||||
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
|
|||||||
@@ -24,7 +24,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_ENDPOINT: http://tei-embedding-service:80
|
TEI_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
interval: 10s
|
interval: 10s
|
||||||
@@ -60,7 +60,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
@@ -96,7 +96,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
HABANA_VISIBLE_DEVICES: all
|
HABANA_VISIBLE_DEVICES: all
|
||||||
|
|||||||
@@ -24,7 +24,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_ENDPOINT: http://tei-embedding-service:80
|
TEI_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
interval: 10s
|
interval: 10s
|
||||||
@@ -60,7 +60,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
@@ -75,7 +75,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
HABANA_VISIBLE_DEVICES: all
|
HABANA_VISIBLE_DEVICES: all
|
||||||
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
|
|||||||
@@ -123,7 +123,7 @@ View the docker input parameters in `./ChatQnA/docker_compose/intel/hpu/gaudi/co
|
|||||||
environment:
|
environment:
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN}
|
||||||
HABANA_VISIBLE_DEVICES: all
|
HABANA_VISIBLE_DEVICES: all
|
||||||
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
||||||
ENABLE_HPU_GRAPH: true
|
ENABLE_HPU_GRAPH: true
|
||||||
|
|||||||
@@ -92,7 +92,7 @@ cat <<EOF > .env
|
|||||||
# Set all required ENV values
|
# Set all required ENV values
|
||||||
export TAG=${TAG}
|
export TAG=${TAG}
|
||||||
export EMBEDDING_MODEL_ID=${EMBEDDING_MODEL_ID}
|
export EMBEDDING_MODEL_ID=${EMBEDDING_MODEL_ID}
|
||||||
export HUGGINGFACEHUB_API_TOKEN=$HUGGINGFACEHUB_API_TOKEN
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export RERANK_MODEL_ID=${RERANK_MODEL_ID}
|
export RERANK_MODEL_ID=${RERANK_MODEL_ID}
|
||||||
export LLM_MODEL_ID=${LLM_MODEL_ID}
|
export LLM_MODEL_ID=${LLM_MODEL_ID}
|
||||||
export INDEX_NAME=${INDEX_NAME}
|
export INDEX_NAME=${INDEX_NAME}
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ pushd "../../../../../" > /dev/null
|
|||||||
source .set_env.sh
|
source .set_env.sh
|
||||||
popd > /dev/null
|
popd > /dev/null
|
||||||
|
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export HF_TOKEN=${HF_TOKEN}
|
export HF_TOKEN=${HF_TOKEN}
|
||||||
export host_ip=${ip_address}
|
export host_ip=${ip_address}
|
||||||
export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
|
export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
|
||||||
|
|||||||
@@ -24,7 +24,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_ENDPOINT: http://tei-embedding-service:80
|
TEI_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
interval: 10s
|
interval: 10s
|
||||||
@@ -76,7 +76,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
||||||
@@ -98,7 +98,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
ipc: host
|
ipc: host
|
||||||
|
|||||||
@@ -12,7 +12,7 @@ export EXTERNAL_HOST_IP=${ip_address}
|
|||||||
export CODEGEN_TGI_SERVICE_PORT=8028
|
export CODEGEN_TGI_SERVICE_PORT=8028
|
||||||
|
|
||||||
### A token for accessing repositories with models
|
### A token for accessing repositories with models
|
||||||
export CODEGEN_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export CODEGEN_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
|
|
||||||
### Model ID
|
### Model ID
|
||||||
export CODEGEN_LLM_MODEL_ID="Qwen/Qwen2.5-Coder-7B-Instruct"
|
export CODEGEN_LLM_MODEL_ID="Qwen/Qwen2.5-Coder-7B-Instruct"
|
||||||
|
|||||||
@@ -13,7 +13,7 @@ export CODEGEN_VLLM_SERVICE_PORT=8028
|
|||||||
export CODEGEN_VLLM_ENDPOINT="http://${HOST_IP}:${CODEGEN_VLLM_SERVICE_PORT}"
|
export CODEGEN_VLLM_ENDPOINT="http://${HOST_IP}:${CODEGEN_VLLM_SERVICE_PORT}"
|
||||||
|
|
||||||
### A token for accessing repositories with models
|
### A token for accessing repositories with models
|
||||||
export CODEGEN_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export CODEGEN_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
|
|
||||||
### Model ID
|
### Model ID
|
||||||
export CODEGEN_LLM_MODEL_ID="Qwen/Qwen2.5-Coder-7B-Instruct"
|
export CODEGEN_LLM_MODEL_ID="Qwen/Qwen2.5-Coder-7B-Instruct"
|
||||||
|
|||||||
@@ -17,7 +17,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
host_ip: ${host_ip}
|
host_ip: ${host_ip}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"]
|
||||||
@@ -39,7 +39,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
host_ip: ${host_ip}
|
host_ip: ${host_ip}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"]
|
||||||
@@ -56,7 +56,7 @@ services:
|
|||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
LLM_ENDPOINT: ${LLM_ENDPOINT}
|
LLM_ENDPOINT: ${LLM_ENDPOINT}
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
llm-tgi-service:
|
llm-tgi-service:
|
||||||
extends: llm-base
|
extends: llm-base
|
||||||
@@ -140,7 +140,7 @@ services:
|
|||||||
REDIS_URL: ${REDIS_URL}
|
REDIS_URL: ${REDIS_URL}
|
||||||
REDIS_HOST: ${host_ip}
|
REDIS_HOST: ${host_ip}
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: true
|
LOGFLAG: true
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
@@ -162,7 +162,7 @@ services:
|
|||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
host_ip: ${host_ip}
|
host_ip: ${host_ip}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD", "curl", "-f", "http://localhost:80/health"]
|
test: ["CMD", "curl", "-f", "http://localhost:80/health"]
|
||||||
interval: 10s
|
interval: 10s
|
||||||
@@ -202,7 +202,7 @@ services:
|
|||||||
REDIS_RETRIEVER_PORT: ${REDIS_RETRIEVER_PORT}
|
REDIS_RETRIEVER_PORT: ${REDIS_RETRIEVER_PORT}
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: ${RETRIEVER_COMPONENT_NAME:-OPEA_RETRIEVER_REDIS}
|
RETRIEVER_COMPONENT_NAME: ${RETRIEVER_COMPONENT_NAME:-OPEA_RETRIEVER_REDIS}
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
|
|||||||
@@ -59,7 +59,7 @@ services:
|
|||||||
REDIS_URL: ${REDIS_URL}
|
REDIS_URL: ${REDIS_URL}
|
||||||
REDIS_HOST: ${host_ip}
|
REDIS_HOST: ${host_ip}
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: true
|
LOGFLAG: true
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
tei-embedding-serving:
|
tei-embedding-serving:
|
||||||
@@ -76,7 +76,7 @@ services:
|
|||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
host_ip: ${host_ip}
|
host_ip: ${host_ip}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD", "curl", "-f", "http://${host_ip}:${TEI_EMBEDDER_PORT}/health"]
|
test: ["CMD", "curl", "-f", "http://${host_ip}:${TEI_EMBEDDER_PORT}/health"]
|
||||||
interval: 10s
|
interval: 10s
|
||||||
@@ -116,7 +116,7 @@ services:
|
|||||||
REDIS_RETRIEVER_PORT: ${REDIS_RETRIEVER_PORT}
|
REDIS_RETRIEVER_PORT: ${REDIS_RETRIEVER_PORT}
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: ${RETRIEVER_COMPONENT_NAME:-OPEA_RETRIEVER_REDIS}
|
RETRIEVER_COMPONENT_NAME: ${RETRIEVER_COMPONENT_NAME:-OPEA_RETRIEVER_REDIS}
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
|
|||||||
@@ -17,7 +17,7 @@ services:
|
|||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HABANA_VISIBLE_DEVICES: all
|
HABANA_VISIBLE_DEVICES: all
|
||||||
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
||||||
HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN}
|
||||||
ENABLE_HPU_GRAPH: true
|
ENABLE_HPU_GRAPH: true
|
||||||
LIMIT_HPU_GRAPH: true
|
LIMIT_HPU_GRAPH: true
|
||||||
USE_FLASH_ATTENTION: true
|
USE_FLASH_ATTENTION: true
|
||||||
@@ -46,7 +46,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
HABANA_VISIBLE_DEVICES: all
|
HABANA_VISIBLE_DEVICES: all
|
||||||
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
||||||
VLLM_SKIP_WARMUP: ${VLLM_SKIP_WARMUP:-false}
|
VLLM_SKIP_WARMUP: ${VLLM_SKIP_WARMUP:-false}
|
||||||
@@ -71,7 +71,7 @@ services:
|
|||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
LLM_ENDPOINT: ${LLM_ENDPOINT}
|
LLM_ENDPOINT: ${LLM_ENDPOINT}
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
llm-tgi-service:
|
llm-tgi-service:
|
||||||
extends: llm-base
|
extends: llm-base
|
||||||
@@ -156,7 +156,7 @@ services:
|
|||||||
REDIS_URL: ${REDIS_URL}
|
REDIS_URL: ${REDIS_URL}
|
||||||
REDIS_HOST: ${host_ip}
|
REDIS_HOST: ${host_ip}
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: true
|
LOGFLAG: true
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
@@ -178,7 +178,7 @@ services:
|
|||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
host_ip: ${host_ip}
|
host_ip: ${host_ip}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD", "curl", "-f", "http://localhost:80/health"]
|
test: ["CMD", "curl", "-f", "http://localhost:80/health"]
|
||||||
interval: 10s
|
interval: 10s
|
||||||
@@ -218,7 +218,7 @@ services:
|
|||||||
REDIS_RETRIEVER_PORT: ${REDIS_RETRIEVER_PORT}
|
REDIS_RETRIEVER_PORT: ${REDIS_RETRIEVER_PORT}
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: ${RETRIEVER_COMPONENT_NAME:-OPEA_RETRIEVER_REDIS}
|
RETRIEVER_COMPONENT_NAME: ${RETRIEVER_COMPONENT_NAME:-OPEA_RETRIEVER_REDIS}
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
|
|||||||
@@ -7,8 +7,8 @@ source .set_env.sh
|
|||||||
popd > /dev/null
|
popd > /dev/null
|
||||||
|
|
||||||
export HOST_IP=$(hostname -I | awk '{print $1}')
|
export HOST_IP=$(hostname -I | awk '{print $1}')
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
if [ -z "${HUGGINGFACEHUB_API_TOKEN}" ]; then
|
if [ -z "${HF_TOKEN}" ]; then
|
||||||
echo "Error: HUGGINGFACEHUB_API_TOKEN is not set. Please set HUGGINGFACEHUB_API_TOKEN"
|
echo "Error: HUGGINGFACEHUB_API_TOKEN is not set. Please set HUGGINGFACEHUB_API_TOKEN"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
|||||||
@@ -21,7 +21,7 @@ export CODETRANS_TGI_SERVICE_PORT=8008
|
|||||||
export CODETRANS_TGI_LLM_ENDPOINT="http://${HOST_IP}:${CODETRANS_TGI_SERVICE_PORT}"
|
export CODETRANS_TGI_LLM_ENDPOINT="http://${HOST_IP}:${CODETRANS_TGI_SERVICE_PORT}"
|
||||||
|
|
||||||
### A token for accessing repositories with models
|
### A token for accessing repositories with models
|
||||||
export CODETRANS_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export CODETRANS_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
|
|
||||||
### The port of the LLM service. On this port, the LLM service will accept connections
|
### The port of the LLM service. On this port, the LLM service will accept connections
|
||||||
export CODETRANS_LLM_SERVICE_PORT=9000
|
export CODETRANS_LLM_SERVICE_PORT=9000
|
||||||
|
|||||||
@@ -21,7 +21,7 @@ export CODETRANS_VLLM_SERVICE_PORT=8008
|
|||||||
export CODETRANS_LLM_ENDPOINT="http://${HOST_IP}:${CODETRANS_VLLM_SERVICE_PORT}"
|
export CODETRANS_LLM_ENDPOINT="http://${HOST_IP}:${CODETRANS_VLLM_SERVICE_PORT}"
|
||||||
|
|
||||||
### A token for accessing repositories with models
|
### A token for accessing repositories with models
|
||||||
export CODETRANS_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export CODETRANS_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
|
|
||||||
### The port of the LLM service. On this port, the LLM service will accept connections
|
### The port of the LLM service. On this port, the LLM service will accept connections
|
||||||
export CODETRANS_LLM_SERVICE_PORT=9000
|
export CODETRANS_LLM_SERVICE_PORT=9000
|
||||||
|
|||||||
@@ -14,7 +14,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
VLLM_TORCH_PROFILER_DIR: "/mnt"
|
VLLM_TORCH_PROFILER_DIR: "/mnt"
|
||||||
healthcheck:
|
healthcheck:
|
||||||
@@ -39,7 +39,7 @@ services:
|
|||||||
LLM_ENDPOINT: ${LLM_ENDPOINT}
|
LLM_ENDPOINT: ${LLM_ENDPOINT}
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
LLM_COMPONENT_NAME: ${LLM_COMPONENT_NAME}
|
LLM_COMPONENT_NAME: ${LLM_COMPONENT_NAME}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
codetrans-xeon-backend-server:
|
codetrans-xeon-backend-server:
|
||||||
image: ${REGISTRY:-opea}/codetrans:${TAG:-latest}
|
image: ${REGISTRY:-opea}/codetrans:${TAG:-latest}
|
||||||
|
|||||||
@@ -14,7 +14,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
host_ip: ${host_ip}
|
host_ip: ${host_ip}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://$host_ip:8008/health || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://$host_ip:8008/health || exit 1"]
|
||||||
@@ -38,7 +38,7 @@ services:
|
|||||||
LLM_ENDPOINT: ${LLM_ENDPOINT}
|
LLM_ENDPOINT: ${LLM_ENDPOINT}
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
LLM_COMPONENT_NAME: ${LLM_COMPONENT_NAME}
|
LLM_COMPONENT_NAME: ${LLM_COMPONENT_NAME}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
codetrans-xeon-backend-server:
|
codetrans-xeon-backend-server:
|
||||||
image: ${REGISTRY:-opea}/codetrans:${TAG:-latest}
|
image: ${REGISTRY:-opea}/codetrans:${TAG:-latest}
|
||||||
|
|||||||
@@ -13,7 +13,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
HABANA_VISIBLE_DEVICES: all
|
HABANA_VISIBLE_DEVICES: all
|
||||||
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
@@ -45,7 +45,7 @@ services:
|
|||||||
LLM_ENDPOINT: ${LLM_ENDPOINT}
|
LLM_ENDPOINT: ${LLM_ENDPOINT}
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
LLM_COMPONENT_NAME: ${LLM_COMPONENT_NAME}
|
LLM_COMPONENT_NAME: ${LLM_COMPONENT_NAME}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
codetrans-gaudi-backend-server:
|
codetrans-gaudi-backend-server:
|
||||||
image: ${REGISTRY:-opea}/codetrans:${TAG:-latest}
|
image: ${REGISTRY:-opea}/codetrans:${TAG:-latest}
|
||||||
|
|||||||
@@ -13,7 +13,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
HABANA_VISIBLE_DEVICES: all
|
HABANA_VISIBLE_DEVICES: all
|
||||||
@@ -42,7 +42,7 @@ services:
|
|||||||
LLM_ENDPOINT: ${LLM_ENDPOINT}
|
LLM_ENDPOINT: ${LLM_ENDPOINT}
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
LLM_COMPONENT_NAME: ${LLM_COMPONENT_NAME}
|
LLM_COMPONENT_NAME: ${LLM_COMPONENT_NAME}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
codetrans-gaudi-backend-server:
|
codetrans-gaudi-backend-server:
|
||||||
image: ${REGISTRY:-opea}/codetrans:${TAG:-latest}
|
image: ${REGISTRY:-opea}/codetrans:${TAG:-latest}
|
||||||
|
|||||||
@@ -38,7 +38,7 @@ function build_docker_images() {
|
|||||||
|
|
||||||
function start_services() {
|
function start_services() {
|
||||||
cd $WORKPATH/docker_compose/intel
|
cd $WORKPATH/docker_compose/intel
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export NGINX_PORT=80
|
export NGINX_PORT=80
|
||||||
export host_ip=${ip_address}
|
export host_ip=${ip_address}
|
||||||
source set_env.sh
|
source set_env.sh
|
||||||
|
|||||||
@@ -40,7 +40,7 @@ function build_docker_images() {
|
|||||||
|
|
||||||
function start_services() {
|
function start_services() {
|
||||||
cd $WORKPATH/docker_compose/intel
|
cd $WORKPATH/docker_compose/intel
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
|
|
||||||
export NGINX_PORT=80
|
export NGINX_PORT=80
|
||||||
export host_ip=${ip_address}
|
export host_ip=${ip_address}
|
||||||
|
|||||||
@@ -35,7 +35,7 @@ function build_docker_images() {
|
|||||||
|
|
||||||
function start_services() {
|
function start_services() {
|
||||||
cd $WORKPATH/docker_compose/intel
|
cd $WORKPATH/docker_compose/intel
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
|
|
||||||
export NGINX_PORT=80
|
export NGINX_PORT=80
|
||||||
export host_ip=${ip_address}
|
export host_ip=${ip_address}
|
||||||
|
|||||||
@@ -35,7 +35,7 @@ function build_docker_images() {
|
|||||||
|
|
||||||
function start_services() {
|
function start_services() {
|
||||||
cd $WORKPATH/docker_compose/intel
|
cd $WORKPATH/docker_compose/intel
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
|
|
||||||
export NGINX_PORT=80
|
export NGINX_PORT=80
|
||||||
export host_ip=${ip_address}
|
export host_ip=${ip_address}
|
||||||
|
|||||||
@@ -9,7 +9,7 @@ source .set_env.sh
|
|||||||
popd > /dev/null
|
popd > /dev/null
|
||||||
|
|
||||||
export host_ip=${ip_address}
|
export host_ip=${ip_address}
|
||||||
export DBQNA_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export DBQNA_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export DBQNA_TGI_SERVICE_PORT=8008
|
export DBQNA_TGI_SERVICE_PORT=8008
|
||||||
export DBQNA_TGI_LLM_ENDPOINT="http://${host_ip}:${DBQNA_TGI_SERVICE_PORT}"
|
export DBQNA_TGI_LLM_ENDPOINT="http://${host_ip}:${DBQNA_TGI_SERVICE_PORT}"
|
||||||
export DBQNA_LLM_MODEL_ID="mistralai/Mistral-7B-Instruct-v0.3"
|
export DBQNA_LLM_MODEL_ID="mistralai/Mistral-7B-Instruct-v0.3"
|
||||||
|
|||||||
@@ -61,7 +61,7 @@ export https_proxy=${https_proxy}
|
|||||||
|
|
||||||
export TGI_PORT=8008
|
export TGI_PORT=8008
|
||||||
export TGI_LLM_ENDPOINT=http://${host_ip}:${TGI_PORT}
|
export TGI_LLM_ENDPOINT=http://${host_ip}:${TGI_PORT}
|
||||||
export HF_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HF_TOKEN=${HF_TOKEN}
|
||||||
export LLM_MODEL_ID="mistralai/Mistral-7B-Instruct-v0.3"
|
export LLM_MODEL_ID="mistralai/Mistral-7B-Instruct-v0.3"
|
||||||
export POSTGRES_USER=postgres
|
export POSTGRES_USER=postgres
|
||||||
export POSTGRES_PASSWORD=testpwd
|
export POSTGRES_PASSWORD=testpwd
|
||||||
@@ -108,7 +108,7 @@ docker run --name test-text2sql-postgres --ipc=host -e POSTGRES_USER=${POSTGRES_
|
|||||||
|
|
||||||
```bash
|
```bash
|
||||||
|
|
||||||
docker run -d --name="test-text2sql-tgi-endpoint" --ipc=host -p $TGI_PORT:80 -v ./data:/data --shm-size 1g -e HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} -e HF_TOKEN=${HF_TOKEN} -e model=${model} ghcr.io/huggingface/text-generation-inference:2.4.1 --model-id $model
|
docker run -d --name="test-text2sql-tgi-endpoint" --ipc=host -p $TGI_PORT:80 -v ./data:/data --shm-size 1g -e HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN} -e HF_TOKEN=${HF_TOKEN} -e model=${model} ghcr.io/huggingface/text-generation-inference:2.4.1 --model-id $model
|
||||||
```
|
```
|
||||||
|
|
||||||
- Start Text-to-SQL Service
|
- Start Text-to-SQL Service
|
||||||
|
|||||||
@@ -13,8 +13,8 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
shm_size: 1g
|
shm_size: 1g
|
||||||
command: --model-id ${LLM_MODEL_ID}
|
command: --model-id ${LLM_MODEL_ID}
|
||||||
|
|
||||||
|
|||||||
@@ -9,8 +9,8 @@ popd > /dev/null
|
|||||||
|
|
||||||
export host_ip=${ip_address}
|
export host_ip=${ip_address}
|
||||||
export no_proxy=$no_proxy,$host_ip,dbqna-xeon-react-ui-server,text2sql-service,tgi-service,postgres-container
|
export no_proxy=$no_proxy,$host_ip,dbqna-xeon-react-ui-server,text2sql-service,tgi-service,postgres-container
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export HF_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HF_TOKEN=${HF_TOKEN}
|
||||||
export POSTGRES_USER=postgres
|
export POSTGRES_USER=postgres
|
||||||
export POSTGRES_PASSWORD=testpwd
|
export POSTGRES_PASSWORD=testpwd
|
||||||
export POSTGRES_DB=chinook
|
export POSTGRES_DB=chinook
|
||||||
|
|||||||
@@ -28,7 +28,7 @@ services:
|
|||||||
REDIS_HOST: ${REDIS_HOST}
|
REDIS_HOST: ${REDIS_HOST}
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
@@ -49,7 +49,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
host_ip: ${host_ip}
|
host_ip: ${host_ip}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"]
|
||||||
@@ -69,7 +69,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
@@ -87,7 +87,7 @@ services:
|
|||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
REDIS_URL: ${REDIS_URL}
|
REDIS_URL: ${REDIS_URL}
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
||||||
@@ -105,7 +105,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
host_ip: ${host_ip}
|
host_ip: ${host_ip}
|
||||||
@@ -129,7 +129,7 @@ services:
|
|||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
RERANK_TYPE: ${RERANK_TYPE}
|
RERANK_TYPE: ${RERANK_TYPE}
|
||||||
TEI_RERANKING_ENDPOINT: ${TEI_RERANKING_ENDPOINT}
|
TEI_RERANKING_ENDPOINT: ${TEI_RERANKING_ENDPOINT}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
|
|||||||
@@ -76,7 +76,7 @@ services:
|
|||||||
DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_MILVUS"
|
DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_MILVUS"
|
||||||
MILVUS_HOST: ${MILVUS_HOST}
|
MILVUS_HOST: ${MILVUS_HOST}
|
||||||
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
@@ -107,7 +107,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
host_ip: ${host_ip}
|
host_ip: ${host_ip}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"]
|
||||||
@@ -130,7 +130,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
@@ -148,7 +148,7 @@ services:
|
|||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
MILVUS_HOST: ${host_ip}
|
MILVUS_HOST: ${host_ip}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_MILVUS"
|
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_MILVUS"
|
||||||
@@ -167,7 +167,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
host_ip: ${host_ip}
|
host_ip: ${host_ip}
|
||||||
@@ -194,7 +194,7 @@ services:
|
|||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
RERANK_TYPE: ${RERANK_TYPE}
|
RERANK_TYPE: ${RERANK_TYPE}
|
||||||
TEI_RERANKING_ENDPOINT: ${TEI_RERANKING_ENDPOINT}
|
TEI_RERANKING_ENDPOINT: ${TEI_RERANKING_ENDPOINT}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
|
|||||||
@@ -25,7 +25,7 @@ services:
|
|||||||
REDIS_HOST: redis-vector-db
|
REDIS_HOST: redis-vector-db
|
||||||
INDEX_NAME: ${INDEX_NAME:-rag-redis}
|
INDEX_NAME: ${INDEX_NAME:-rag-redis}
|
||||||
TEI_ENDPOINT: http://tei-embedding-service:80
|
TEI_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
@@ -46,7 +46,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
host_ip: ${host_ip}
|
host_ip: ${host_ip}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"]
|
||||||
@@ -66,7 +66,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
@@ -84,7 +84,7 @@ services:
|
|||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
REDIS_URL: redis://redis-vector-db:6379
|
REDIS_URL: redis://redis-vector-db:6379
|
||||||
INDEX_NAME: ${INDEX_NAME:-rag-redis}
|
INDEX_NAME: ${INDEX_NAME:-rag-redis}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
||||||
|
|||||||
@@ -12,7 +12,7 @@ export RERANK_MODEL_ID="BAAI/bge-reranker-base"
|
|||||||
export TEI_EMBEDDING_ENDPOINT="http://${ip_address}:6006"
|
export TEI_EMBEDDING_ENDPOINT="http://${ip_address}:6006"
|
||||||
export TEI_RERANKING_ENDPOINT="http://${ip_address}:8808"
|
export TEI_RERANKING_ENDPOINT="http://${ip_address}:8808"
|
||||||
export TGI_LLM_ENDPOINT="http://${ip_address}:8008"
|
export TGI_LLM_ENDPOINT="http://${ip_address}:8008"
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export MEGA_SERVICE_HOST_IP=${ip_address}
|
export MEGA_SERVICE_HOST_IP=${ip_address}
|
||||||
export EMBEDDING_SERVICE_HOST_IP=${ip_address}
|
export EMBEDDING_SERVICE_HOST_IP=${ip_address}
|
||||||
export RETRIEVER_SERVICE_HOST_IP=${ip_address}
|
export RETRIEVER_SERVICE_HOST_IP=${ip_address}
|
||||||
|
|||||||
@@ -28,7 +28,7 @@ services:
|
|||||||
REDIS_URL: ${REDIS_URL}
|
REDIS_URL: ${REDIS_URL}
|
||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
interval: 10s
|
interval: 10s
|
||||||
@@ -76,7 +76,7 @@ services:
|
|||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
retriever:
|
retriever:
|
||||||
@@ -96,7 +96,7 @@ services:
|
|||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
||||||
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
tei-reranking-service:
|
tei-reranking-service:
|
||||||
image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.6
|
image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.6
|
||||||
@@ -111,7 +111,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
host_ip: ${host_ip}
|
host_ip: ${host_ip}
|
||||||
@@ -135,7 +135,7 @@ services:
|
|||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
RERANK_TYPE: ${RERANK_TYPE}
|
RERANK_TYPE: ${RERANK_TYPE}
|
||||||
TEI_RERANKING_ENDPOINT: ${TEI_RERANKING_ENDPOINT}
|
TEI_RERANKING_ENDPOINT: ${TEI_RERANKING_ENDPOINT}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
|
|||||||
@@ -76,7 +76,7 @@ services:
|
|||||||
DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_MILVUS"
|
DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_MILVUS"
|
||||||
MILVUS_HOST: ${MILVUS_HOST}
|
MILVUS_HOST: ${MILVUS_HOST}
|
||||||
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
|
||||||
@@ -136,7 +136,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
@@ -154,7 +154,7 @@ services:
|
|||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
MILVUS_HOST: ${host_ip}
|
MILVUS_HOST: ${host_ip}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_MILVUS"
|
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_MILVUS"
|
||||||
@@ -173,7 +173,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
host_ip: ${host_ip}
|
host_ip: ${host_ip}
|
||||||
@@ -200,7 +200,7 @@ services:
|
|||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
RERANK_TYPE: ${RERANK_TYPE}
|
RERANK_TYPE: ${RERANK_TYPE}
|
||||||
TEI_RERANKING_ENDPOINT: ${TEI_RERANKING_ENDPOINT}
|
TEI_RERANKING_ENDPOINT: ${TEI_RERANKING_ENDPOINT}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
|
|||||||
@@ -11,7 +11,7 @@ export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
|
|||||||
export RERANK_MODEL_ID="BAAI/bge-reranker-base"
|
export RERANK_MODEL_ID="BAAI/bge-reranker-base"
|
||||||
export TEI_EMBEDDING_ENDPOINT="http://${ip_address}:8090"
|
export TEI_EMBEDDING_ENDPOINT="http://${ip_address}:8090"
|
||||||
export TEI_RERANKING_ENDPOINT="http://${ip_address}:8808"
|
export TEI_RERANKING_ENDPOINT="http://${ip_address}:8808"
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export MEGA_SERVICE_HOST_IP=${ip_address}
|
export MEGA_SERVICE_HOST_IP=${ip_address}
|
||||||
export EMBEDDING_SERVICE_HOST_IP=${ip_address}
|
export EMBEDDING_SERVICE_HOST_IP=${ip_address}
|
||||||
export RETRIEVER_SERVICE_HOST_IP=${ip_address}
|
export RETRIEVER_SERVICE_HOST_IP=${ip_address}
|
||||||
|
|||||||
@@ -43,7 +43,7 @@ function start_services() {
|
|||||||
export TEI_EMBEDDING_ENDPOINT="http://${ip_address}:6006"
|
export TEI_EMBEDDING_ENDPOINT="http://${ip_address}:6006"
|
||||||
export REDIS_URL="redis://${ip_address}:6379"
|
export REDIS_URL="redis://${ip_address}:6379"
|
||||||
export INDEX_NAME="rag-redis"
|
export INDEX_NAME="rag-redis"
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export MEGA_SERVICE_HOST_IP=${ip_address}
|
export MEGA_SERVICE_HOST_IP=${ip_address}
|
||||||
export EMBEDDING_SERVICE_HOST_IP=${ip_address}
|
export EMBEDDING_SERVICE_HOST_IP=${ip_address}
|
||||||
export RETRIEVER_SERVICE_HOST_IP=${ip_address}
|
export RETRIEVER_SERVICE_HOST_IP=${ip_address}
|
||||||
|
|||||||
@@ -9,7 +9,7 @@ export DOCSUM_MAX_TOTAL_TOKENS="4096"
|
|||||||
export DOCSUM_LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
|
export DOCSUM_LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
|
||||||
export DOCSUM_TGI_SERVICE_PORT="8008"
|
export DOCSUM_TGI_SERVICE_PORT="8008"
|
||||||
export DOCSUM_TGI_LLM_ENDPOINT="http://${HOST_IP}:${DOCSUM_TGI_SERVICE_PORT}"
|
export DOCSUM_TGI_LLM_ENDPOINT="http://${HOST_IP}:${DOCSUM_TGI_SERVICE_PORT}"
|
||||||
export DOCSUM_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export DOCSUM_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export DOCSUM_WHISPER_PORT="7066"
|
export DOCSUM_WHISPER_PORT="7066"
|
||||||
export ASR_SERVICE_HOST_IP="${HOST_IP}"
|
export ASR_SERVICE_HOST_IP="${HOST_IP}"
|
||||||
export DOCSUM_LLM_SERVER_PORT="9000"
|
export DOCSUM_LLM_SERVER_PORT="9000"
|
||||||
|
|||||||
@@ -4,7 +4,7 @@
|
|||||||
# SPDX-License-Identifier: Apache-2.0
|
# SPDX-License-Identifier: Apache-2.0
|
||||||
|
|
||||||
export HOST_IP=${ip_address}
|
export HOST_IP=${ip_address}
|
||||||
export DOCSUM_HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export DOCSUM_HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
export DOCSUM_MAX_INPUT_TOKENS=2048
|
export DOCSUM_MAX_INPUT_TOKENS=2048
|
||||||
export DOCSUM_MAX_TOTAL_TOKENS=4096
|
export DOCSUM_MAX_TOTAL_TOKENS=4096
|
||||||
export DOCSUM_LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
|
export DOCSUM_LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
|
||||||
|
|||||||
@@ -14,7 +14,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
VLLM_TORCH_PROFILER_DIR: "/mnt"
|
VLLM_TORCH_PROFILER_DIR: "/mnt"
|
||||||
healthcheck:
|
healthcheck:
|
||||||
@@ -39,7 +39,7 @@ services:
|
|||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
LLM_ENDPOINT: ${LLM_ENDPOINT}
|
LLM_ENDPOINT: ${LLM_ENDPOINT}
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
MAX_INPUT_TOKENS: ${MAX_INPUT_TOKENS}
|
MAX_INPUT_TOKENS: ${MAX_INPUT_TOKENS}
|
||||||
MAX_TOTAL_TOKENS: ${MAX_TOTAL_TOKENS}
|
MAX_TOTAL_TOKENS: ${MAX_TOTAL_TOKENS}
|
||||||
DocSum_COMPONENT_NAME: ${DocSum_COMPONENT_NAME}
|
DocSum_COMPONENT_NAME: ${DocSum_COMPONENT_NAME}
|
||||||
|
|||||||
@@ -14,7 +14,7 @@ services:
|
|||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
TGI_LLM_ENDPOINT: ${TGI_LLM_ENDPOINT}
|
TGI_LLM_ENDPOINT: ${TGI_LLM_ENDPOINT}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
host_ip: ${host_ip}
|
host_ip: ${host_ip}
|
||||||
healthcheck:
|
healthcheck:
|
||||||
test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"]
|
test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"]
|
||||||
@@ -39,7 +39,7 @@ services:
|
|||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
LLM_ENDPOINT: ${LLM_ENDPOINT}
|
LLM_ENDPOINT: ${LLM_ENDPOINT}
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
MAX_INPUT_TOKENS: ${MAX_INPUT_TOKENS}
|
MAX_INPUT_TOKENS: ${MAX_INPUT_TOKENS}
|
||||||
MAX_TOTAL_TOKENS: ${MAX_TOTAL_TOKENS}
|
MAX_TOTAL_TOKENS: ${MAX_TOTAL_TOKENS}
|
||||||
DocSum_COMPONENT_NAME: ${DocSum_COMPONENT_NAME}
|
DocSum_COMPONENT_NAME: ${DocSum_COMPONENT_NAME}
|
||||||
|
|||||||
@@ -13,7 +13,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HF_TOKEN: ${HF_TOKEN}
|
||||||
HABANA_VISIBLE_DEVICES: all
|
HABANA_VISIBLE_DEVICES: all
|
||||||
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
||||||
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
LLM_MODEL_ID: ${LLM_MODEL_ID}
|
||||||
@@ -44,7 +44,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
MAX_INPUT_TOKENS: ${MAX_INPUT_TOKENS}
|
MAX_INPUT_TOKENS: ${MAX_INPUT_TOKENS}
|
||||||
MAX_TOTAL_TOKENS: ${MAX_TOTAL_TOKENS}
|
MAX_TOTAL_TOKENS: ${MAX_TOTAL_TOKENS}
|
||||||
LLM_ENDPOINT: ${LLM_ENDPOINT}
|
LLM_ENDPOINT: ${LLM_ENDPOINT}
|
||||||
|
|||||||
@@ -13,7 +13,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN}
|
||||||
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
HF_HUB_DISABLE_PROGRESS_BARS: 1
|
||||||
HF_HUB_ENABLE_HF_TRANSFER: 0
|
HF_HUB_ENABLE_HF_TRANSFER: 0
|
||||||
HABANA_VISIBLE_DEVICES: all
|
HABANA_VISIBLE_DEVICES: all
|
||||||
@@ -48,7 +48,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN}
|
||||||
MAX_INPUT_TOKENS: ${MAX_INPUT_TOKENS}
|
MAX_INPUT_TOKENS: ${MAX_INPUT_TOKENS}
|
||||||
MAX_TOTAL_TOKENS: ${MAX_TOTAL_TOKENS}
|
MAX_TOTAL_TOKENS: ${MAX_TOTAL_TOKENS}
|
||||||
LLM_ENDPOINT: ${LLM_ENDPOINT}
|
LLM_ENDPOINT: ${LLM_ENDPOINT}
|
||||||
|
|||||||
@@ -10,7 +10,7 @@ export host_ip=$(hostname -I | awk '{print $1}') # Example: host_ip="192.168.1.1
|
|||||||
export no_proxy="${no_proxy},${host_ip}" # Example: no_proxy="localhost, 127.0.0.1, 192.168.1.1"
|
export no_proxy="${no_proxy},${host_ip}" # Example: no_proxy="localhost, 127.0.0.1, 192.168.1.1"
|
||||||
export http_proxy=$http_proxy
|
export http_proxy=$http_proxy
|
||||||
export https_proxy=$https_proxy
|
export https_proxy=$https_proxy
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HF_TOKEN}
|
||||||
|
|
||||||
export LLM_ENDPOINT_PORT=8008
|
export LLM_ENDPOINT_PORT=8008
|
||||||
export LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
|
export LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
|
||||||
|
|||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user