Enable health check for dataprep in ChatQnA (#1799)

Signed-off-by: letonghan <letong.han@intel.com>
This commit is contained in:
Letong Han
2025-04-17 15:01:57 +08:00
committed by GitHub
parent 4fc19c7d73
commit ae31e4fb75
20 changed files with 369 additions and 113 deletions

View File

@@ -32,6 +32,12 @@ services:
INDEX_NAME: ${INDEX_NAME}
TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s
timeout: 5s
retries: 50
restart: unless-stopped
tei-embedding-service:
image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.6
container_name: tei-embedding-server
@@ -107,11 +113,18 @@ services:
image: ${REGISTRY:-opea}/chatqna:${TAG:-latest}
container_name: chatqna-xeon-backend-server
depends_on:
- redis-vector-db
- tei-embedding-service
- retriever
- tei-reranking-service
- vllm-service
redis-vector-db:
condition: service_started
dataprep-redis-service:
condition: service_healthy
tei-embedding-service:
condition: service_started
retriever:
condition: service_started
tei-reranking-service:
condition: service_started
vllm-service:
condition: service_healthy
ports:
- "8888:8888"
environment:

View File

@@ -25,6 +25,12 @@ services:
INDEX_NAME: ${INDEX_NAME}
TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s
timeout: 5s
retries: 50
restart: unless-stopped
tei-embedding-service:
image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.6
container_name: tei-embedding-server
@@ -121,12 +127,20 @@ services:
image: ${REGISTRY:-opea}/chatqna:${TAG:-latest}
container_name: chatqna-xeon-backend-server
depends_on:
- redis-vector-db
- tei-embedding-service
- retriever
- tei-reranking-service
- vllm-service
- llm-faqgen
redis-vector-db:
condition: service_started
tei-embedding-service:
condition: service_started
retriever:
condition: service_started
tei-reranking-service:
condition: service_started
vllm-service:
condition: service_started
llm-faqgen:
condition: service_started
dataprep-redis-service:
condition: service_healthy
ports:
- ${CHATQNA_BACKEND_PORT:-8888}:8888
environment:

View File

@@ -25,6 +25,12 @@ services:
INDEX_NAME: ${INDEX_NAME}
TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s
timeout: 5s
retries: 50
restart: unless-stopped
tei-embedding-service:
image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.6
container_name: tei-embedding-server
@@ -121,12 +127,20 @@ services:
image: ${REGISTRY:-opea}/chatqna:${TAG:-latest}
container_name: chatqna-xeon-backend-server
depends_on:
- redis-vector-db
- tei-embedding-service
- retriever
- tei-reranking-service
- tgi-service
- llm-faqgen
redis-vector-db:
condition: service_started
tei-embedding-service:
condition: service_started
retriever:
condition: service_started
tei-reranking-service:
condition: service_started
tgi-service:
condition: service_started
llm-faqgen:
condition: service_started
dataprep-redis-service:
condition: service_healthy
ports:
- ${CHATQNA_BACKEND_PORT:-8888}:8888
environment:

View File

@@ -78,6 +78,11 @@ services:
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
EMBEDDING_MODEL_ID: ${EMBEDDING_MODEL_ID}
LOGFLAG: ${LOGFLAG}
healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s
timeout: 5s
retries: 50
restart: unless-stopped
depends_on:
milvus-standalone:
@@ -154,7 +159,7 @@ services:
LLM_MODEL_ID: ${LLM_MODEL_ID}
VLLM_TORCH_PROFILER_DIR: "/mnt"
healthcheck:
test: ["CMD-SHELL", "curl -f http://$host_ip:9009/health || exit 1"]
test: ["CMD-SHELL", "curl -f http://localhost:80/health || exit 1"]
interval: 10s
timeout: 10s
retries: 100
@@ -164,12 +169,16 @@ services:
image: ${REGISTRY:-opea}/chatqna:${TAG:-latest}
container_name: chatqna-xeon-backend-server
depends_on:
- milvus-standalone
- tei-embedding-service
- dataprep-milvus-service
- retriever
- tei-reranking-service
- vllm-service
tei-embedding-service:
condition: service_started
dataprep-milvus-service:
condition: service_healthy
retriever:
condition: service_started
tei-reranking-service:
condition: service_started
vllm-service:
condition: service_healthy
ports:
- "8888:8888"
environment:

View File

@@ -22,6 +22,12 @@ services:
LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_PINECONE"
healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s
timeout: 5s
retries: 50
restart: unless-stopped
tei-embedding-service:
image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.6
container_name: tei-embedding-server
@@ -89,11 +95,16 @@ services:
image: ${REGISTRY:-opea}/chatqna:${TAG:-latest}
container_name: chatqna-xeon-backend-server
depends_on:
- tei-embedding-service
- dataprep-pinecone-service
- retriever
- tei-reranking-service
- vllm-service
tei-embedding-service:
condition: service_started
dataprep-pinecone-service:
condition: service_healthy
retriever:
condition: service_started
tei-reranking-service:
condition: service_started
vllm-service:
condition: service_started
ports:
- "8888:8888"
environment:

View File

@@ -26,6 +26,12 @@ services:
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_QDRANT"
healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s
timeout: 5s
retries: 50
restart: unless-stopped
tei-embedding-service:
image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5
container_name: tei-embedding-server
@@ -94,11 +100,18 @@ services:
image: ${REGISTRY:-opea}/chatqna:${TAG:-latest}
container_name: chatqna-xeon-backend-server
depends_on:
- qdrant-vector-db
- tei-embedding-service
- retriever
- tei-reranking-service
- vllm-service
qdrant-vector-db:
condition: service_started
dataprep-qdrant-service:
condition: service_healthy
tei-embedding-service:
condition: service_started
retriever:
condition: service_started
tei-reranking-service:
condition: service_started
vllm-service:
condition: service_started
ports:
- "8912:8888"
environment:

View File

@@ -25,6 +25,12 @@ services:
INDEX_NAME: ${INDEX_NAME}
TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s
timeout: 5s
retries: 50
restart: unless-stopped
tei-embedding-service:
image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.6
container_name: tei-embedding-server
@@ -94,12 +100,18 @@ services:
image: ${REGISTRY:-opea}/chatqna:${TAG:-latest}
container_name: chatqna-xeon-backend-server
depends_on:
- redis-vector-db
- tei-embedding-service
- dataprep-redis-service
- retriever
- tei-reranking-service
- tgi-service
redis-vector-db:
condition: service_started
tei-embedding-service:
condition: service_started
dataprep-redis-service:
condition: service_healthy
retriever:
condition: service_started
tei-reranking-service:
condition: service_started
tgi-service:
condition: service_started
ports:
- "8888:8888"
environment:

View File

@@ -25,6 +25,12 @@ services:
INDEX_NAME: ${INDEX_NAME}
TEI_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
healthcheck:
test: ["CMD-SHELL", "curl -f http://localhost:5000/v1/health_check || exit 1"]
interval: 10s
timeout: 5s
retries: 50
restart: unless-stopped
tei-embedding-service:
image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.6
container_name: tei-embedding-server
@@ -78,11 +84,16 @@ services:
image: ${REGISTRY:-opea}/chatqna:${TAG:-latest}
container_name: chatqna-xeon-backend-server
depends_on:
- redis-vector-db
- tei-embedding-service
- dataprep-redis-service
- retriever
- vllm-service
redis-vector-db:
condition: service_started
tei-embedding-service:
condition: service_started
dataprep-redis-service:
condition: service_healthy
retriever:
condition: service_started
vllm-service:
condition: service_started
ports:
- "8888:8888"
environment: