Fix README issues (#817)

Signed-off-by: lvliang-intel <liang1.lv@intel.com>
Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
This commit is contained in:
lvliang-intel
2024-09-18 09:50:17 +08:00
committed by GitHub
parent 375ea7a90c
commit bceacdc804
24 changed files with 106 additions and 243 deletions

View File

@@ -14,20 +14,15 @@ After launching your instance, you can connect to it using SSH (for Linux instan
First of all, you need to build Docker Images locally and install the python package of it. This step can be ignored after the Docker images published to Docker hub.
### 1. Install GenAIComps from Source Code
### 1. Build the LLM Docker Image
```bash
git clone https://github.com/opea-project/GenAIComps.git
cd GenAIComps
```
### 2. Build the LLM Docker Image
```bash
docker build -t opea/llm-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile .
```
### 3. Build MegaService Docker Image
### 2. Build MegaService Docker Image
```bash
git clone https://github.com/opea-project/GenAIExamples.git
@@ -35,14 +30,14 @@ cd GenAIExamples/CodeTrans
docker build -t opea/codetrans:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile .
```
### 4. Build UI Docker Image
### 3. Build UI Docker Image
```bash
cd GenAIExamples/CodeTrans/ui
docker build -t opea/codetrans-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile .
```
### 5. Build Nginx Docker Image
### 4. Build Nginx Docker Image
```bash
cd GenAIComps