rename docker compose.yaml (#446)

Signed-off-by: chensuyue <suyue.chen@intel.com>
This commit is contained in:
chen, suyue
2024-07-26 20:03:36 +08:00
committed by GitHub
parent 8f9bcd4b3c
commit 965c13c556
72 changed files with 389 additions and 387 deletions

View File

@@ -15,9 +15,9 @@
version: "3.8"
services:
tgi_service:
tgi-service:
image: ghcr.io/huggingface/tgi-gaudi:2.0.1
container_name: tgi_gaudi_service
container_name: tgi-gaudi-server
ports:
- "8008:80"
environment:
@@ -32,7 +32,7 @@ services:
image: opea/llm-tgi:latest
container_name: llm-tgi-gaudi-server
depends_on:
- tgi_service
- tgi-service
ports:
- "9000:9000"
ipc: host
@@ -46,7 +46,7 @@ services:
image: opea/translation:latest
container_name: translation-gaudi-backend-server
depends_on:
- tgi_service
- tgi-service
- llm
ports:
- "8888:8888"

View File

@@ -15,9 +15,9 @@
version: "3.8"
services:
tgi_service:
tgi-service:
image: ghcr.io/huggingface/text-generation-inference:1.4
container_name: tgi_service
container_name: tgi-service
ports:
- "8008:80"
environment:
@@ -33,7 +33,7 @@ services:
image: opea/llm-tgi:latest
container_name: llm-tgi-server
depends_on:
- tgi_service
- tgi-service
ports:
- "9000:9000"
ipc: host
@@ -50,7 +50,7 @@ services:
image: opea/translation:latest
container_name: translation-xeon-backend-server
depends_on:
- tgi_service
- tgi-service
- llm
ports:
- "8888:8888"

View File

@@ -50,7 +50,15 @@ function start_services() {
# Start Docker Containers
docker compose up -d
sleep 2m # Waits 2 minutes
n=0
until [[ "$n" -ge 500 ]]; do
docker logs tgi-gaudi-server > ${LOG_PATH}/tgi_service_start.log
if grep -q Connected ${LOG_PATH}/tgi_service_start.log; then
break
fi
sleep 1s
n=$((n+1))
done
}
function validate_services() {
@@ -92,7 +100,7 @@ function validate_microservices() {
"${ip_address}:8008/generate" \
"generated_text" \
"tgi-gaudi" \
"tgi_gaudi_service" \
"tgi-gaudi-service" \
'{"inputs":"What is Deep Learning?","parameters":{"max_new_tokens":17, "do_sample": true}}'
# llm microservice
@@ -141,7 +149,7 @@ function validate_frontend() {
function stop_docker() {
cd $WORKPATH/docker/gaudi
docker compose down
docker compose stop && docker compose rm -f
}
function main() {

View File

@@ -49,7 +49,15 @@ function start_services() {
# Start Docker Containers
docker compose up -d
sleep 2m # Waits 2 minutes
n=0
until [[ "$n" -ge 500 ]]; do
docker logs tgi-service > ${LOG_PATH}/tgi_service_start.log
if grep -q Connected ${LOG_PATH}/tgi_service_start.log; then
break
fi
sleep 1s
n=$((n+1))
done
}
function validate_services() {
@@ -89,7 +97,7 @@ function validate_microservices() {
"${ip_address}:8008/generate" \
"generated_text" \
"tgi" \
"tgi_service" \
"tgi-service" \
'{"inputs":"What is Deep Learning?","parameters":{"max_new_tokens":17, "do_sample": true}}'
# llm microservice
@@ -138,7 +146,7 @@ function validate_frontend() {
function stop_docker() {
cd $WORKPATH/docker/xeon
docker compose down
docker compose stop && docker compose rm -f
}
function main() {

View File

@@ -3,7 +3,7 @@
# SPDX-License-Identifier: Apache-2.0
opea_micro_services:
tgi_service:
tgi-service:
host: ${TGI_SERVICE_IP}
ports: ${TGI_SERVICE_PORT}
image: ghcr.io/huggingface/tgi-gaudi:2.0.1