Check duplicated dockerfile (#1289)
Signed-off-by: ZePan110 <ze.pan@intel.com>
This commit is contained in:
@@ -11,7 +11,7 @@ First of all, you need to build Docker Images locally and install the python pac
|
||||
```bash
|
||||
git clone https://github.com/opea-project/GenAIComps.git
|
||||
cd GenAIComps
|
||||
docker build -t opea/llm-tgi:latest --no-cache --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/src/text-generation/Dockerfile .
|
||||
docker build -t opea/llm-textgen:latest --no-cache --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/src/text-generation/Dockerfile .
|
||||
```
|
||||
|
||||
### 2. Build MegaService Docker Image
|
||||
@@ -38,7 +38,7 @@ docker build -t opea/nginx:latest --build-arg https_proxy=$https_proxy --build-a
|
||||
|
||||
Then run the command `docker images`, you will have the following Docker Images:
|
||||
|
||||
- `opea/llm-tgi:latest`
|
||||
- `opea/llm-textgen:latest`
|
||||
- `opea/codetrans:latest`
|
||||
- `opea/codetrans-ui:latest`
|
||||
- `opea/nginx:latest`
|
||||
|
||||
@@ -31,8 +31,8 @@ services:
|
||||
ipc: host
|
||||
command: --model-id ${LLM_MODEL_ID} --max-input-length 1024 --max-total-tokens 2048
|
||||
llm:
|
||||
image: ${REGISTRY:-opea}/llm-tgi:${TAG:-latest}
|
||||
container_name: llm-tgi-gaudi-server
|
||||
image: ${REGISTRY:-opea}/llm-textgen:${TAG:-latest}
|
||||
container_name: llm-textgen-gaudi-server
|
||||
depends_on:
|
||||
tgi-service:
|
||||
condition: service_healthy
|
||||
|
||||
Reference in New Issue
Block a user