# Copyright (C) 2024 Advanced Micro Devices, Inc. # SPDX-License-Identifier: Apache-2.0 services: docsum-tgi-service: image: ghcr.io/huggingface/text-generation-inference:2.4.1-rocm container_name: docsum-tgi-service ports: - "${DOCSUM_TGI_SERVICE_PORT:-8008}:80" environment: no_proxy: ${no_proxy} http_proxy: ${http_proxy} https_proxy: ${https_proxy} TGI_LLM_ENDPOINT: "http://${HOST_IP}:${DOCSUM_TGI_SERVICE_PORT}" HUGGINGFACEHUB_API_TOKEN: ${DOCSUM_HUGGINGFACEHUB_API_TOKEN} host_ip: ${host_ip} DOCSUM_TGI_SERVICE_PORT: ${DOCSUM_TGI_SERVICE_PORT} volumes: - "${MODEL_CACHE:-./data}:/data" shm_size: 20g devices: - /dev/kfd:/dev/kfd - /dev/dri/:/dev/dri/ cap_add: - SYS_PTRACE group_add: - video security_opt: - seccomp:unconfined ipc: host healthcheck: test: ["CMD-SHELL", "curl -f http://${host_ip}:${DOCSUM_TGI_SERVICE_PORT}/health || exit 1"] interval: 10s timeout: 10s retries: 100 command: --model-id ${DOCSUM_LLM_MODEL_ID} --max-input-length ${DOCSUM_MAX_INPUT_TOKENS} --max-total-tokens ${DOCSUM_MAX_TOTAL_TOKENS} docsum-llm-server: image: ${REGISTRY:-opea}/llm-docsum:${TAG:-latest} container_name: docsum-llm-server depends_on: docsum-tgi-service: condition: service_healthy ports: - "${DOCSUM_LLM_SERVER_PORT}:9000" ipc: host environment: no_proxy: ${no_proxy} http_proxy: ${http_proxy} https_proxy: ${https_proxy} LLM_ENDPOINT: ${DOCSUM_TGI_LLM_ENDPOINT} HUGGINGFACEHUB_API_TOKEN: ${DOCSUM_HUGGINGFACEHUB_API_TOKEN} MAX_INPUT_TOKENS: ${DOCSUM_MAX_INPUT_TOKENS} MAX_TOTAL_TOKENS: ${DOCSUM_MAX_TOTAL_TOKENS} LLM_MODEL_ID: ${DOCSUM_LLM_MODEL_ID} DocSum_COMPONENT_NAME: "OpeaDocSumTgi" LOGFLAG: ${LOGFLAG:-False} restart: unless-stopped whisper: image: ${REGISTRY:-opea}/whisper:${TAG:-latest} container_name: whisper-service ports: - "${DOCSUM_WHISPER_PORT:-7066}:7066" ipc: host environment: no_proxy: ${no_proxy} http_proxy: ${http_proxy} https_proxy: ${https_proxy} restart: unless-stopped docsum-backend-server: image: ${REGISTRY:-opea}/docsum:${TAG:-latest} container_name: docsum-backend-server depends_on: - docsum-tgi-service - docsum-llm-server ports: - "${DOCSUM_BACKEND_SERVER_PORT}:8888" environment: no_proxy: ${no_proxy} https_proxy: ${https_proxy} http_proxy: ${http_proxy} MEGA_SERVICE_HOST_IP: ${HOST_IP} LLM_SERVICE_HOST_IP: ${HOST_IP} LLM_SERVICE_PORT: ${DOCSUM_LLM_SERVER_PORT} ASR_SERVICE_HOST_IP: ${ASR_SERVICE_HOST_IP} ASR_SERVICE_PORT: ${DOCSUM_WHISPER_PORT} ipc: host restart: always docsum-gradio-ui: image: ${REGISTRY:-opea}/docsum-gradio-ui:${TAG:-latest} container_name: docsum-ui-server depends_on: - docsum-backend-server ports: - "5173:5173" environment: no_proxy: ${no_proxy} https_proxy: ${https_proxy} http_proxy: ${http_proxy} BACKEND_SERVICE_ENDPOINT: ${BACKEND_SERVICE_ENDPOINT} DOC_BASE_URL: ${BACKEND_SERVICE_ENDPOINT} ipc: host restart: always networks: default: driver: bridge