Files
GenAIExamples/CodeGen/codegen.yaml
2024-05-30 16:04:59 +08:00

54 lines
1.5 KiB
YAML

# Copyright (c) 2024 Intel Corporation
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
opea_micro_services:
tgi_service:
host: ${TGI_SERVICE_IP}
ports: ${TGI_SERVICE_PORT}
image: ghcr.io/huggingface/tgi-gaudi:1.2.1
volumes:
- "./data:/data"
runtime: habana
cap_add:
- SYS_NICE
ipc: host
environment:
HF_TOKEN: ${HF_TOKEN}
HABANA_VISIBLE_DEVICES: all
OMPI_MCA_btl_vader_single_copy_mechanism: none
model-id: ${LLM_MODEL_ID}
llm:
host: ${LLM_SERVICE_HOST_IP}
ports: ${LLM_SERVICE_PORT}
image: opea/llm-tgi:latest
endpoint: /v1/chat/completions
environment:
TGI_LLM_ENDPOINT: ${TGI_LLM_ENDPOINT}
HF_TOKEN: ${HF_TOKEN}
ui:
host: ${UI_SERVICE_HOST_IP}
ports:
- "5173:5173"
image: opea/codegen-ui:latest
environment:
- CHAT_BASE_URL=${BACKEND_SERVICE_ENDPOINT}
opea_mega_service:
host: ${MEGA_SERVICE_HOST_IP}
ports: ${MEGA_SERVICE_PORT}
endpoint: /v1/codegen
image: opea/codegen:latest
mega_flow:
- llm