Compare commits
1 Commits
main
...
revert-148
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
2836f11716 |
@@ -91,14 +91,6 @@ cd GenAIExamples/ChatQnA/docker_compose/intel/cpu/xeon/
|
|||||||
docker compose up -d
|
docker compose up -d
|
||||||
```
|
```
|
||||||
|
|
||||||
To enable Open Telemetry Tracing, compose.telemetry.yaml file need to be merged along with default compose.yaml file.
|
|
||||||
CPU example with Open Telemetry feature:
|
|
||||||
|
|
||||||
```bash
|
|
||||||
cd GenAIExamples/ChatQnA/docker_compose/intel/cpu/xeon/
|
|
||||||
docker compose -f compose.yaml -f compose.telemetry.yaml up -d
|
|
||||||
```
|
|
||||||
|
|
||||||
It will automatically download the docker image on `docker hub`:
|
It will automatically download the docker image on `docker hub`:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
@@ -240,13 +232,6 @@ cd GenAIExamples/ChatQnA/docker_compose/intel/hpu/gaudi/
|
|||||||
docker compose up -d
|
docker compose up -d
|
||||||
```
|
```
|
||||||
|
|
||||||
To enable Open Telemetry Tracing, compose.telemetry.yaml file need to be merged along with default compose.yaml file.
|
|
||||||
|
|
||||||
```bash
|
|
||||||
cd GenAIExamples/ChatQnA/docker_compose/intel/hpu/gaudi/
|
|
||||||
docker compose -f compose.yaml -f compose.telemetry.yaml up -d
|
|
||||||
```
|
|
||||||
|
|
||||||
Refer to the [Gaudi Guide](./docker_compose/intel/hpu/gaudi/README.md) to build docker images from source.
|
Refer to the [Gaudi Guide](./docker_compose/intel/hpu/gaudi/README.md) to build docker images from source.
|
||||||
|
|
||||||
### Deploy ChatQnA on Xeon
|
### Deploy ChatQnA on Xeon
|
||||||
@@ -258,13 +243,6 @@ cd GenAIExamples/ChatQnA/docker_compose/intel/cpu/xeon/
|
|||||||
docker compose up -d
|
docker compose up -d
|
||||||
```
|
```
|
||||||
|
|
||||||
To enable Open Telemetry Tracing, compose.telemetry.yaml file need to be merged along with default compose.yaml file.
|
|
||||||
|
|
||||||
```bash
|
|
||||||
cd GenAIExamples/ChatQnA/docker_compose/intel/cpu/xeon/
|
|
||||||
docker compose -f compose.yaml -f compose.telemetry.yaml up -d
|
|
||||||
```
|
|
||||||
|
|
||||||
Refer to the [Xeon Guide](./docker_compose/intel/cpu/xeon/README.md) for more instructions on building docker images from source.
|
Refer to the [Xeon Guide](./docker_compose/intel/cpu/xeon/README.md) for more instructions on building docker images from source.
|
||||||
|
|
||||||
### Deploy ChatQnA on NVIDIA GPU
|
### Deploy ChatQnA on NVIDIA GPU
|
||||||
@@ -368,7 +346,7 @@ OPEA microservice deployment can easily be monitored through Grafana dashboards
|
|||||||
|
|
||||||
## Tracing Services with OpenTelemetry Tracing and Jaeger
|
## Tracing Services with OpenTelemetry Tracing and Jaeger
|
||||||
|
|
||||||
> NOTE: This feature is disabled by default. Please check the Deploy ChatQnA sessions for how to enable this feature with compose_telemetry.yaml file.
|
> NOTE: limited support. Only LLM inference serving with TGI on Gaudi is enabled for this feature.
|
||||||
|
|
||||||
OPEA microservice and TGI/TEI serving can easily be traced through Jaeger dashboards in conjunction with OpenTelemetry Tracing feature. Follow the [README](https://github.com/opea-project/GenAIComps/tree/main/comps/cores/telemetry#tracing) to trace additional functions if needed.
|
OPEA microservice and TGI/TEI serving can easily be traced through Jaeger dashboards in conjunction with OpenTelemetry Tracing feature. Follow the [README](https://github.com/opea-project/GenAIComps/tree/main/comps/cores/telemetry#tracing) to trace additional functions if needed.
|
||||||
|
|
||||||
|
|||||||
@@ -44,14 +44,6 @@ To set up environment variables for deploying ChatQnA services, follow these ste
|
|||||||
docker compose up -d
|
docker compose up -d
|
||||||
```
|
```
|
||||||
|
|
||||||
To enable Open Telemetry Tracing, compose.telemetry.yaml file need to be merged along with default compose.yaml file.
|
|
||||||
CPU example with Open Telemetry feature:
|
|
||||||
|
|
||||||
```bash
|
|
||||||
cd GenAIExamples/ChatQnA/docker_compose/intel/cpu/xeon/
|
|
||||||
docker compose -f compose.yaml -f compose.telemetry.yaml up -d
|
|
||||||
```
|
|
||||||
|
|
||||||
It will automatically download the docker image on `docker hub`:
|
It will automatically download the docker image on `docker hub`:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
@@ -271,16 +263,12 @@ If use vLLM as the LLM serving backend.
|
|||||||
docker compose -f compose.yaml up -d
|
docker compose -f compose.yaml up -d
|
||||||
# Start ChatQnA without Rerank Pipeline
|
# Start ChatQnA without Rerank Pipeline
|
||||||
docker compose -f compose_without_rerank.yaml up -d
|
docker compose -f compose_without_rerank.yaml up -d
|
||||||
# Start ChatQnA with Rerank Pipeline and Open Telemetry Tracing
|
|
||||||
docker compose -f compose.yaml -f compose.telemetry.yaml up -d
|
|
||||||
```
|
```
|
||||||
|
|
||||||
If use TGI as the LLM serving backend.
|
If use TGI as the LLM serving backend.
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
docker compose -f compose_tgi.yaml up -d
|
docker compose -f compose_tgi.yaml up -d
|
||||||
# Start ChatQnA with Open Telemetry Tracing
|
|
||||||
docker compose -f compose_tgi.yaml -f compose_tgi.telemetry.yaml up -d
|
|
||||||
```
|
```
|
||||||
|
|
||||||
### Validate Microservices
|
### Validate Microservices
|
||||||
|
|||||||
@@ -1,27 +0,0 @@
|
|||||||
# Copyright (C) 2024 Intel Corporation
|
|
||||||
# SPDX-License-Identifier: Apache-2.0
|
|
||||||
|
|
||||||
services:
|
|
||||||
tei-embedding-service:
|
|
||||||
command: --model-id ${EMBEDDING_MODEL_ID} --auto-truncate --otlp-endpoint $OTEL_EXPORTER_OTLP_TRACES_ENDPOINT
|
|
||||||
tei-reranking-service:
|
|
||||||
command: --model-id ${RERANK_MODEL_ID} --auto-truncate --otlp-endpoint $OTEL_EXPORTER_OTLP_TRACES_ENDPOINT
|
|
||||||
jaeger:
|
|
||||||
image: jaegertracing/all-in-one:latest
|
|
||||||
container_name: jaeger
|
|
||||||
ports:
|
|
||||||
- "16686:16686"
|
|
||||||
- "4317:4317"
|
|
||||||
- "4318:4318"
|
|
||||||
- "9411:9411"
|
|
||||||
ipc: host
|
|
||||||
environment:
|
|
||||||
no_proxy: ${no_proxy}
|
|
||||||
http_proxy: ${http_proxy}
|
|
||||||
https_proxy: ${https_proxy}
|
|
||||||
COLLECTOR_ZIPKIN_HOST_PORT: 9411
|
|
||||||
restart: unless-stopped
|
|
||||||
chatqna-xeon-backend-server:
|
|
||||||
environment:
|
|
||||||
- ENABLE_OPEA_TELEMETRY=true
|
|
||||||
- TELEMETRY_ENDPOINT=${TELEMETRY_ENDPOINT}
|
|
||||||
@@ -1,29 +0,0 @@
|
|||||||
# Copyright (C) 2024 Intel Corporation
|
|
||||||
# SPDX-License-Identifier: Apache-2.0
|
|
||||||
|
|
||||||
services:
|
|
||||||
tei-embedding-service:
|
|
||||||
command: --model-id ${EMBEDDING_MODEL_ID} --auto-truncate --otlp-endpoint $OTEL_EXPORTER_OTLP_TRACES_ENDPOINT
|
|
||||||
tei-reranking-service:
|
|
||||||
command: --model-id ${RERANK_MODEL_ID} --auto-truncate --otlp-endpoint $OTEL_EXPORTER_OTLP_TRACES_ENDPOINT
|
|
||||||
tgi-service:
|
|
||||||
command: --model-id ${LLM_MODEL_ID} --cuda-graphs 0 --otlp-endpoint $OTEL_EXPORTER_OTLP_TRACES_ENDPOINT
|
|
||||||
jaeger:
|
|
||||||
image: jaegertracing/all-in-one:latest
|
|
||||||
container_name: jaeger
|
|
||||||
ports:
|
|
||||||
- "16686:16686"
|
|
||||||
- "4317:4317"
|
|
||||||
- "4318:4318"
|
|
||||||
- "9411:9411"
|
|
||||||
ipc: host
|
|
||||||
environment:
|
|
||||||
no_proxy: ${no_proxy}
|
|
||||||
http_proxy: ${http_proxy}
|
|
||||||
https_proxy: ${https_proxy}
|
|
||||||
COLLECTOR_ZIPKIN_HOST_PORT: 9411
|
|
||||||
restart: unless-stopped
|
|
||||||
chatqna-xeon-backend-server:
|
|
||||||
environment:
|
|
||||||
- ENABLE_OPEA_TELEMETRY=true
|
|
||||||
- TELEMETRY_ENDPOINT=${TELEMETRY_ENDPOINT}
|
|
||||||
@@ -19,7 +19,3 @@ export INDEX_NAME="rag-redis"
|
|||||||
# Set it as a non-null string, such as true, if you want to enable logging facility,
|
# Set it as a non-null string, such as true, if you want to enable logging facility,
|
||||||
# otherwise, keep it as "" to disable it.
|
# otherwise, keep it as "" to disable it.
|
||||||
export LOGFLAG=""
|
export LOGFLAG=""
|
||||||
# Set OpenTelemetry Tracing Endpoint
|
|
||||||
export JAEGER_IP=$(ip route get 8.8.8.8 | grep -oP 'src \K[^ ]+')
|
|
||||||
export OTEL_EXPORTER_OTLP_TRACES_ENDPOINT=grpc://$JAEGER_IP:4317
|
|
||||||
export TELEMETRY_ENDPOINT=http://$JAEGER_IP:4318/v1/traces
|
|
||||||
|
|||||||
@@ -45,12 +45,6 @@ To set up environment variables for deploying ChatQnA services, follow these ste
|
|||||||
docker compose up -d
|
docker compose up -d
|
||||||
```
|
```
|
||||||
|
|
||||||
To enable Open Telemetry Tracing, compose.telemetry.yaml file need to be merged along with default compose.yaml file.
|
|
||||||
|
|
||||||
```bash
|
|
||||||
docker compose -f compose.yaml -f compose.telemetry.yaml up -d
|
|
||||||
```
|
|
||||||
|
|
||||||
It will automatically download the docker image on `docker hub`:
|
It will automatically download the docker image on `docker hub`:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
@@ -265,16 +259,12 @@ If use vLLM as the LLM serving backend.
|
|||||||
docker compose -f compose.yaml up -d
|
docker compose -f compose.yaml up -d
|
||||||
# Start ChatQnA without Rerank Pipeline
|
# Start ChatQnA without Rerank Pipeline
|
||||||
docker compose -f compose_without_rerank.yaml up -d
|
docker compose -f compose_without_rerank.yaml up -d
|
||||||
# Start ChatQnA with Rerank Pipeline and Open Telemetry Tracing
|
|
||||||
docker compose -f compose.yaml -f compose.telemetry.yaml up -d
|
|
||||||
```
|
```
|
||||||
|
|
||||||
If use TGI as the LLM serving backend.
|
If use TGI as the LLM serving backend.
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
docker compose -f compose_tgi.yaml up -d
|
docker compose -f compose_tgi.yaml up -d
|
||||||
# Start ChatQnA with Open Telemetry Tracing
|
|
||||||
docker compose -f compose_tgi.yaml -f compose_tgi.telemetry.yaml up -d
|
|
||||||
```
|
```
|
||||||
|
|
||||||
If you want to enable guardrails microservice in the pipeline, please follow the below command instead:
|
If you want to enable guardrails microservice in the pipeline, please follow the below command instead:
|
||||||
|
|||||||
@@ -1,27 +0,0 @@
|
|||||||
# Copyright (C) 2024 Intel Corporation
|
|
||||||
# SPDX-License-Identifier: Apache-2.0
|
|
||||||
|
|
||||||
services:
|
|
||||||
tei-embedding-service:
|
|
||||||
command: --model-id ${EMBEDDING_MODEL_ID} --auto-truncate --otlp-endpoint $OTEL_EXPORTER_OTLP_TRACES_ENDPOINT
|
|
||||||
tei-reranking-service:
|
|
||||||
command: --model-id ${RERANK_MODEL_ID} --auto-truncate --otlp-endpoint $OTEL_EXPORTER_OTLP_TRACES_ENDPOINT
|
|
||||||
jaeger:
|
|
||||||
image: jaegertracing/all-in-one:latest
|
|
||||||
container_name: jaeger
|
|
||||||
ports:
|
|
||||||
- "16686:16686"
|
|
||||||
- "4317:4317"
|
|
||||||
- "4318:4318"
|
|
||||||
- "9411:9411"
|
|
||||||
ipc: host
|
|
||||||
environment:
|
|
||||||
no_proxy: ${no_proxy}
|
|
||||||
http_proxy: ${http_proxy}
|
|
||||||
https_proxy: ${https_proxy}
|
|
||||||
COLLECTOR_ZIPKIN_HOST_PORT: 9411
|
|
||||||
restart: unless-stopped
|
|
||||||
chatqna-gaudi-backend-server:
|
|
||||||
environment:
|
|
||||||
- ENABLE_OPEA_TELEMETRY=true
|
|
||||||
- TELEMETRY_ENDPOINT=${TELEMETRY_ENDPOINT}
|
|
||||||
@@ -1,29 +0,0 @@
|
|||||||
# Copyright (C) 2024 Intel Corporation
|
|
||||||
# SPDX-License-Identifier: Apache-2.0
|
|
||||||
|
|
||||||
services:
|
|
||||||
tei-embedding-service:
|
|
||||||
command: --model-id ${EMBEDDING_MODEL_ID} --auto-truncate --otlp-endpoint $OTEL_EXPORTER_OTLP_TRACES_ENDPOINT
|
|
||||||
tei-reranking-service:
|
|
||||||
command: --model-id ${RERANK_MODEL_ID} --auto-truncate --otlp-endpoint $OTEL_EXPORTER_OTLP_TRACES_ENDPOINT
|
|
||||||
tgi-service:
|
|
||||||
command: --model-id ${LLM_MODEL_ID} --max-input-length 2048 --max-total-tokens 4096 --otlp-endpoint $OTEL_EXPORTER_OTLP_TRACES_ENDPOINT
|
|
||||||
jaeger:
|
|
||||||
image: jaegertracing/all-in-one:latest
|
|
||||||
container_name: jaeger
|
|
||||||
ports:
|
|
||||||
- "16686:16686"
|
|
||||||
- "4317:4317"
|
|
||||||
- "4318:4318"
|
|
||||||
- "9411:9411"
|
|
||||||
ipc: host
|
|
||||||
environment:
|
|
||||||
no_proxy: ${no_proxy}
|
|
||||||
http_proxy: ${http_proxy}
|
|
||||||
https_proxy: ${https_proxy}
|
|
||||||
COLLECTOR_ZIPKIN_HOST_PORT: 9411
|
|
||||||
restart: unless-stopped
|
|
||||||
chatqna-gaudi-backend-server:
|
|
||||||
environment:
|
|
||||||
- ENABLE_OPEA_TELEMETRY=true
|
|
||||||
- TELEMETRY_ENDPOINT=${TELEMETRY_ENDPOINT}
|
|
||||||
@@ -25,6 +25,7 @@ services:
|
|||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_ENDPOINT: http://tei-embedding-service:80
|
TEI_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
||||||
|
TELEMETRY_ENDPOINT: ${TELEMETRY_ENDPOINT}
|
||||||
tei-embedding-service:
|
tei-embedding-service:
|
||||||
image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5
|
image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5
|
||||||
container_name: tei-embedding-gaudi-server
|
container_name: tei-embedding-gaudi-server
|
||||||
@@ -37,7 +38,7 @@ services:
|
|||||||
no_proxy: ${no_proxy}
|
no_proxy: ${no_proxy}
|
||||||
http_proxy: ${http_proxy}
|
http_proxy: ${http_proxy}
|
||||||
https_proxy: ${https_proxy}
|
https_proxy: ${https_proxy}
|
||||||
command: --model-id ${EMBEDDING_MODEL_ID} --auto-truncate
|
command: --model-id ${EMBEDDING_MODEL_ID} --auto-truncate --otlp-endpoint $OTEL_EXPORTER_OTLP_TRACES_ENDPOINT
|
||||||
retriever:
|
retriever:
|
||||||
image: ${REGISTRY:-opea}/retriever:${TAG:-latest}
|
image: ${REGISTRY:-opea}/retriever:${TAG:-latest}
|
||||||
container_name: retriever-redis-server
|
container_name: retriever-redis-server
|
||||||
@@ -55,6 +56,7 @@ services:
|
|||||||
INDEX_NAME: ${INDEX_NAME}
|
INDEX_NAME: ${INDEX_NAME}
|
||||||
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
|
||||||
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
|
||||||
|
TELEMETRY_ENDPOINT: ${TELEMETRY_ENDPOINT}
|
||||||
LOGFLAG: ${LOGFLAG}
|
LOGFLAG: ${LOGFLAG}
|
||||||
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
@@ -78,7 +80,7 @@ services:
|
|||||||
HABANA_VISIBLE_DEVICES: all
|
HABANA_VISIBLE_DEVICES: all
|
||||||
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
OMPI_MCA_btl_vader_single_copy_mechanism: none
|
||||||
MAX_WARMUP_SEQUENCE_LENGTH: 512
|
MAX_WARMUP_SEQUENCE_LENGTH: 512
|
||||||
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
|
command: --model-id ${RERANK_MODEL_ID} --auto-truncate --otlp-endpoint $OTEL_EXPORTER_OTLP_TRACES_ENDPOINT
|
||||||
tgi-service:
|
tgi-service:
|
||||||
image: ghcr.io/huggingface/tgi-gaudi:2.0.6
|
image: ghcr.io/huggingface/tgi-gaudi:2.0.6
|
||||||
container_name: tgi-gaudi-server
|
container_name: tgi-gaudi-server
|
||||||
@@ -104,7 +106,22 @@ services:
|
|||||||
cap_add:
|
cap_add:
|
||||||
- SYS_NICE
|
- SYS_NICE
|
||||||
ipc: host
|
ipc: host
|
||||||
command: --model-id ${LLM_MODEL_ID} --num-shard ${NUM_CARDS} --max-input-length 2048 --max-total-tokens 4096
|
command: --model-id ${LLM_MODEL_ID} --num-shard ${NUM_CARDS} --max-input-length 2048 --max-total-tokens 4096 --otlp-endpoint $OTEL_EXPORTER_OTLP_TRACES_ENDPOINT
|
||||||
|
jaeger:
|
||||||
|
image: jaegertracing/all-in-one:latest
|
||||||
|
container_name: jaeger
|
||||||
|
ports:
|
||||||
|
- "16686:16686"
|
||||||
|
- "4317:4317"
|
||||||
|
- "4318:4318"
|
||||||
|
- "9411:9411"
|
||||||
|
ipc: host
|
||||||
|
environment:
|
||||||
|
no_proxy: ${no_proxy}
|
||||||
|
http_proxy: ${http_proxy}
|
||||||
|
https_proxy: ${https_proxy}
|
||||||
|
COLLECTOR_ZIPKIN_HOST_PORT: 9411
|
||||||
|
restart: unless-stopped
|
||||||
chatqna-gaudi-backend-server:
|
chatqna-gaudi-backend-server:
|
||||||
image: ${REGISTRY:-opea}/chatqna:${TAG:-latest}
|
image: ${REGISTRY:-opea}/chatqna:${TAG:-latest}
|
||||||
container_name: chatqna-gaudi-backend-server
|
container_name: chatqna-gaudi-backend-server
|
||||||
@@ -130,6 +147,7 @@ services:
|
|||||||
- LLM_SERVER_PORT=${LLM_SERVER_PORT:-80}
|
- LLM_SERVER_PORT=${LLM_SERVER_PORT:-80}
|
||||||
- LLM_MODEL=${LLM_MODEL_ID}
|
- LLM_MODEL=${LLM_MODEL_ID}
|
||||||
- LOGFLAG=${LOGFLAG}
|
- LOGFLAG=${LOGFLAG}
|
||||||
|
- TELEMETRY_ENDPOINT=${TELEMETRY_ENDPOINT}
|
||||||
ipc: host
|
ipc: host
|
||||||
restart: always
|
restart: always
|
||||||
chatqna-gaudi-ui-server:
|
chatqna-gaudi-ui-server:
|
||||||
|
|||||||
@@ -49,12 +49,9 @@ function start_services() {
|
|||||||
export INDEX_NAME="rag-redis"
|
export INDEX_NAME="rag-redis"
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
||||||
export host_ip=${ip_address}
|
export host_ip=${ip_address}
|
||||||
export JAEGER_IP=$(ip route get 8.8.8.8 | grep -oP 'src \K[^ ]+')
|
|
||||||
export OTEL_EXPORTER_OTLP_TRACES_ENDPOINT=grpc://$JAEGER_IP:4317
|
|
||||||
export TELEMETRY_ENDPOINT=http://$JAEGER_IP:4318/v1/traces
|
|
||||||
|
|
||||||
# Start Docker Containers
|
# Start Docker Containers
|
||||||
docker compose -f compose.yaml -f compose_telemetry.yaml up -d > ${LOG_PATH}/start_services_with_compose.log
|
docker compose -f compose.yaml up -d > ${LOG_PATH}/start_services_with_compose.log
|
||||||
n=0
|
n=0
|
||||||
until [[ "$n" -ge 160 ]]; do
|
until [[ "$n" -ge 160 ]]; do
|
||||||
echo "n=$n"
|
echo "n=$n"
|
||||||
@@ -175,7 +172,7 @@ function validate_frontend() {
|
|||||||
|
|
||||||
function stop_docker() {
|
function stop_docker() {
|
||||||
cd $WORKPATH/docker_compose/intel/hpu/gaudi
|
cd $WORKPATH/docker_compose/intel/hpu/gaudi
|
||||||
docker compose -f compose.yaml -f compose_telemetry.yaml down
|
docker compose -f compose.yaml down
|
||||||
}
|
}
|
||||||
|
|
||||||
function main() {
|
function main() {
|
||||||
|
|||||||
@@ -49,12 +49,9 @@ function start_services() {
|
|||||||
export INDEX_NAME="rag-redis"
|
export INDEX_NAME="rag-redis"
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
||||||
export host_ip=${ip_address}
|
export host_ip=${ip_address}
|
||||||
export JAEGER_IP=$(ip route get 8.8.8.8 | grep -oP 'src \K[^ ]+')
|
|
||||||
export OTEL_EXPORTER_OTLP_TRACES_ENDPOINT=grpc://$JAEGER_IP:4317
|
|
||||||
export TELEMETRY_ENDPOINT=http://$JAEGER_IP:4318/v1/traces
|
|
||||||
|
|
||||||
# Start Docker Containers
|
# Start Docker Containers
|
||||||
docker compose -f compose.yaml -f compose_telemetry.yaml up -d > ${LOG_PATH}/start_services_with_compose.log
|
docker compose -f compose.yaml up -d > ${LOG_PATH}/start_services_with_compose.log
|
||||||
n=0
|
n=0
|
||||||
until [[ "$n" -ge 100 ]]; do
|
until [[ "$n" -ge 100 ]]; do
|
||||||
docker logs vllm-service > ${LOG_PATH}/vllm_service_start.log 2>&1
|
docker logs vllm-service > ${LOG_PATH}/vllm_service_start.log 2>&1
|
||||||
@@ -175,7 +172,7 @@ function validate_frontend() {
|
|||||||
|
|
||||||
function stop_docker() {
|
function stop_docker() {
|
||||||
cd $WORKPATH/docker_compose/intel/cpu/xeon
|
cd $WORKPATH/docker_compose/intel/cpu/xeon
|
||||||
docker compose -f compose.yaml -f compose_telemetry.yaml down
|
docker compose -f compose.yaml down
|
||||||
}
|
}
|
||||||
|
|
||||||
function main() {
|
function main() {
|
||||||
|
|||||||
@@ -54,7 +54,7 @@ function start_services() {
|
|||||||
export TELEMETRY_ENDPOINT=http://$JAEGER_IP:4318/v1/traces
|
export TELEMETRY_ENDPOINT=http://$JAEGER_IP:4318/v1/traces
|
||||||
|
|
||||||
# Start Docker Containers
|
# Start Docker Containers
|
||||||
docker compose -f compose_tgi.yaml -f compose_tgi_telemetry.yaml up -d > ${LOG_PATH}/start_services_with_compose.log
|
docker compose -f compose_tgi.yaml up -d > ${LOG_PATH}/start_services_with_compose.log
|
||||||
|
|
||||||
n=0
|
n=0
|
||||||
until [[ "$n" -ge 500 ]]; do
|
until [[ "$n" -ge 500 ]]; do
|
||||||
@@ -218,7 +218,7 @@ function validate_frontend() {
|
|||||||
|
|
||||||
function stop_docker() {
|
function stop_docker() {
|
||||||
cd $WORKPATH/docker_compose/intel/hpu/gaudi
|
cd $WORKPATH/docker_compose/intel/hpu/gaudi
|
||||||
docker compose -f compose_tgi.yaml -f compose_tgi_telemetry.yaml down
|
docker compose -f compose_tgi.yaml down
|
||||||
}
|
}
|
||||||
|
|
||||||
function main() {
|
function main() {
|
||||||
|
|||||||
@@ -48,12 +48,9 @@ function start_services() {
|
|||||||
export LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
|
export LLM_MODEL_ID="meta-llama/Meta-Llama-3-8B-Instruct"
|
||||||
export INDEX_NAME="rag-redis"
|
export INDEX_NAME="rag-redis"
|
||||||
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN}
|
||||||
export JAEGER_IP=$(ip route get 8.8.8.8 | grep -oP 'src \K[^ ]+')
|
|
||||||
export OTEL_EXPORTER_OTLP_TRACES_ENDPOINT=grpc://$JAEGER_IP:4317
|
|
||||||
export TELEMETRY_ENDPOINT=http://$JAEGER_IP:4318/v1/traces
|
|
||||||
|
|
||||||
# Start Docker Containers
|
# Start Docker Containers
|
||||||
docker compose -f compose_tgi.yaml -f compose_tgi_telemetry.yaml up -d > ${LOG_PATH}/start_services_with_compose.log
|
docker compose -f compose_tgi.yaml up -d > ${LOG_PATH}/start_services_with_compose.log
|
||||||
|
|
||||||
n=0
|
n=0
|
||||||
until [[ "$n" -ge 100 ]]; do
|
until [[ "$n" -ge 100 ]]; do
|
||||||
@@ -219,7 +216,7 @@ function validate_frontend() {
|
|||||||
|
|
||||||
function stop_docker() {
|
function stop_docker() {
|
||||||
cd $WORKPATH/docker_compose/intel/cpu/xeon
|
cd $WORKPATH/docker_compose/intel/cpu/xeon
|
||||||
docker compose -f compose_tgi.yaml -f compose_tgi_telemetry.yaml down
|
docker compose -f compose_tgi.yaml down
|
||||||
}
|
}
|
||||||
|
|
||||||
function main() {
|
function main() {
|
||||||
|
|||||||
Reference in New Issue
Block a user