rename docker compose.yaml (#446)
Signed-off-by: chensuyue <suyue.chen@intel.com>
This commit is contained in:
@@ -15,9 +15,9 @@
|
||||
version: "3.8"
|
||||
|
||||
services:
|
||||
tgi_service:
|
||||
tgi-service:
|
||||
image: ghcr.io/huggingface/tgi-gaudi:2.0.1
|
||||
container_name: tgi_gaudi_service
|
||||
container_name: tgi-gaudi-server
|
||||
ports:
|
||||
- "8008:80"
|
||||
environment:
|
||||
@@ -32,7 +32,7 @@ services:
|
||||
image: opea/llm-tgi:latest
|
||||
container_name: llm-tgi-gaudi-server
|
||||
depends_on:
|
||||
- tgi_service
|
||||
- tgi-service
|
||||
ports:
|
||||
- "9000:9000"
|
||||
ipc: host
|
||||
@@ -46,7 +46,7 @@ services:
|
||||
image: opea/translation:latest
|
||||
container_name: translation-gaudi-backend-server
|
||||
depends_on:
|
||||
- tgi_service
|
||||
- tgi-service
|
||||
- llm
|
||||
ports:
|
||||
- "8888:8888"
|
||||
|
||||
@@ -15,9 +15,9 @@
|
||||
version: "3.8"
|
||||
|
||||
services:
|
||||
tgi_service:
|
||||
tgi-service:
|
||||
image: ghcr.io/huggingface/text-generation-inference:1.4
|
||||
container_name: tgi_service
|
||||
container_name: tgi-service
|
||||
ports:
|
||||
- "8008:80"
|
||||
environment:
|
||||
@@ -33,7 +33,7 @@ services:
|
||||
image: opea/llm-tgi:latest
|
||||
container_name: llm-tgi-server
|
||||
depends_on:
|
||||
- tgi_service
|
||||
- tgi-service
|
||||
ports:
|
||||
- "9000:9000"
|
||||
ipc: host
|
||||
@@ -50,7 +50,7 @@ services:
|
||||
image: opea/translation:latest
|
||||
container_name: translation-xeon-backend-server
|
||||
depends_on:
|
||||
- tgi_service
|
||||
- tgi-service
|
||||
- llm
|
||||
ports:
|
||||
- "8888:8888"
|
||||
|
||||
@@ -50,7 +50,15 @@ function start_services() {
|
||||
# Start Docker Containers
|
||||
docker compose up -d
|
||||
|
||||
sleep 2m # Waits 2 minutes
|
||||
n=0
|
||||
until [[ "$n" -ge 500 ]]; do
|
||||
docker logs tgi-gaudi-server > ${LOG_PATH}/tgi_service_start.log
|
||||
if grep -q Connected ${LOG_PATH}/tgi_service_start.log; then
|
||||
break
|
||||
fi
|
||||
sleep 1s
|
||||
n=$((n+1))
|
||||
done
|
||||
}
|
||||
|
||||
function validate_services() {
|
||||
@@ -92,7 +100,7 @@ function validate_microservices() {
|
||||
"${ip_address}:8008/generate" \
|
||||
"generated_text" \
|
||||
"tgi-gaudi" \
|
||||
"tgi_gaudi_service" \
|
||||
"tgi-gaudi-service" \
|
||||
'{"inputs":"What is Deep Learning?","parameters":{"max_new_tokens":17, "do_sample": true}}'
|
||||
|
||||
# llm microservice
|
||||
@@ -141,7 +149,7 @@ function validate_frontend() {
|
||||
|
||||
function stop_docker() {
|
||||
cd $WORKPATH/docker/gaudi
|
||||
docker compose down
|
||||
docker compose stop && docker compose rm -f
|
||||
}
|
||||
|
||||
function main() {
|
||||
|
||||
@@ -49,7 +49,15 @@ function start_services() {
|
||||
# Start Docker Containers
|
||||
docker compose up -d
|
||||
|
||||
sleep 2m # Waits 2 minutes
|
||||
n=0
|
||||
until [[ "$n" -ge 500 ]]; do
|
||||
docker logs tgi-service > ${LOG_PATH}/tgi_service_start.log
|
||||
if grep -q Connected ${LOG_PATH}/tgi_service_start.log; then
|
||||
break
|
||||
fi
|
||||
sleep 1s
|
||||
n=$((n+1))
|
||||
done
|
||||
}
|
||||
|
||||
function validate_services() {
|
||||
@@ -89,7 +97,7 @@ function validate_microservices() {
|
||||
"${ip_address}:8008/generate" \
|
||||
"generated_text" \
|
||||
"tgi" \
|
||||
"tgi_service" \
|
||||
"tgi-service" \
|
||||
'{"inputs":"What is Deep Learning?","parameters":{"max_new_tokens":17, "do_sample": true}}'
|
||||
|
||||
# llm microservice
|
||||
@@ -138,7 +146,7 @@ function validate_frontend() {
|
||||
|
||||
function stop_docker() {
|
||||
cd $WORKPATH/docker/xeon
|
||||
docker compose down
|
||||
docker compose stop && docker compose rm -f
|
||||
}
|
||||
|
||||
function main() {
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
# SPDX-License-Identifier: Apache-2.0
|
||||
|
||||
opea_micro_services:
|
||||
tgi_service:
|
||||
tgi-service:
|
||||
host: ${TGI_SERVICE_IP}
|
||||
ports: ${TGI_SERVICE_PORT}
|
||||
image: ghcr.io/huggingface/tgi-gaudi:2.0.1
|
||||
|
||||
Reference in New Issue
Block a user