Refactor lvm related examples (#1333)

This commit is contained in:
Sihan Chen
2025-01-13 13:42:06 +08:00
committed by GitHub
parent f48bd8e74f
commit ca15fe9bdb
25 changed files with 162 additions and 155 deletions

View File

@@ -71,10 +71,10 @@ docker build -t opea/reranking:latest --build-arg https_proxy=$https_proxy --bui
### 4. Build LVM Image (Xeon)
```bash
docker build -t opea/video-llama-lvm-server:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/lvms/video-llama/dependency/Dockerfile .
docker build -t opea/lvm-video-llama:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/lvms/src/integrations/dependency/video-llama/Dockerfile .
# LVM Service Image
docker build -t opea/lvm-video-llama:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/lvms/video-llama/Dockerfile .
docker build -t opea/lvm:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/lvms/src/Dockerfile .
```
### 5. Build Dataprep Image
@@ -109,11 +109,16 @@ Then run the command `docker images`, you will have the following 8 Docker Image
1. `opea/dataprep-multimodal-vdms:latest`
2. `opea/embedding-multimodal-clip:latest`
3. `opea/retriever-vdms:latest`
<<<<<<< HEAD
4. `opea/reranking:latest`
5. `opea/video-llama-lvm-server:latest`
6. `opea/lvm-video-llama:latest`
7. `opea/videoqna:latest`
8. `opea/videoqna-ui:latest`
6. # `opea/lvm-video-llama:latest`
7. `opea/reranking-tei:latest`
8. `opea/lvm-video-llama:latest`
9. `opea/lvm:latest`
> > > > > > > d93597cbfd9da92b956adb3673c9e5d743c181af
10. `opea/videoqna:latest`
11. `opea/videoqna-ui:latest`
## 🚀 Start Microservices
@@ -275,7 +280,7 @@ docker compose up -d
In first startup, this service will take times to download the LLM file. After it's finished, the service will be ready.
Use `docker logs video-llama-lvm-server` to check if the download is finished.
Use `docker logs lvm-video-llama` to check if the download is finished.
```bash
curl -X POST \

View File

@@ -75,8 +75,8 @@ services:
DATAPREP_GET_VIDEO_LIST_ENDPOINT: ${DATAPREP_GET_VIDEO_LIST_ENDPOINT}
restart: unless-stopped
lvm-video-llama:
image: ${REGISTRY:-opea}/video-llama-lvm-server:${TAG:-latest}
container_name: video-llama-lvm-server
image: ${REGISTRY:-opea}/lvm-video-llama:${TAG:-latest}
container_name: lvm-video-llama
ports:
- "9009:9009"
ipc: host
@@ -90,8 +90,8 @@ services:
- video-llama-model:/home/user/model
restart: unless-stopped
lvm:
image: ${REGISTRY:-opea}/lvm-video-llama:${TAG:-latest}
container_name: lvm-video-llama
image: ${REGISTRY:-opea}/lvm:${TAG:-latest}
container_name: lvm
ports:
- "9000:9000"
ipc: host
@@ -99,6 +99,7 @@ services:
http_proxy: ${http_proxy}
https_proxy: ${https_proxy}
no_proxy: ${no_proxy}
LVM_COMPONENT_NAME: "OPEA_VIDEO_LLAMA_LVM"
LVM_ENDPOINT: ${LVM_ENDPOINT}
restart: unless-stopped
depends_on: