Compare commits

...

4 Commits

Author SHA1 Message Date
CICD-at-OPEA
c712c01829 Update vLLM-fork version to v0.7.2+Gaudi-1.21.0
Signed-off-by: CICD-at-OPEA <CICD@opea.dev>
2025-05-20 22:41:44 +00:00
chen, suyue
66ecaf2e6f Merge branch 'main' into update_vLLM-fork 2025-05-20 10:54:35 +08:00
xiguiw
46fab4a736 Merge branch 'main' into update_vLLM-fork 2025-05-16 09:34:11 +08:00
CICD-at-OPEA
8ba40a5bff Update vLLM-fork version to v0.7.2+Gaudi-1.21.0
Signed-off-by: CICD-at-OPEA <CICD@opea.dev>
2025-05-08 08:37:52 +00:00
12 changed files with 12 additions and 12 deletions

View File

@@ -2,4 +2,4 @@
# SPDX-License-Identifier: Apache-2.0
export VLLM_VER=v0.8.3
export VLLM_FORK_VER=v0.6.6.post1+Gaudi-1.20.0
export VLLM_FORK_VER=v0.7.2+Gaudi-1.21.0

View File

@@ -37,7 +37,7 @@ function build_agent_docker_image_gaudi_vllm() {
get_genai_comps
git clone https://github.com/HabanaAI/vllm-fork.git && cd vllm-fork
VLLM_FORK_VER=v0.6.6.post1+Gaudi-1.20.0
VLLM_FORK_VER=v0.7.2+Gaudi-1.21.0
git checkout ${VLLM_FORK_VER} &> /dev/null && cd ../
echo "Build agent image with --no-cache..."

View File

@@ -27,7 +27,7 @@ function build_docker_images() {
git clone https://github.com/HabanaAI/vllm-fork.git
cd vllm-fork/
VLLM_FORK_VER=v0.6.6.post1+Gaudi-1.20.0
VLLM_FORK_VER=v0.7.2+Gaudi-1.21.0
echo "Check out vLLM tag ${VLLM_FORK_VER}"
git checkout ${VLLM_FORK_VER} &> /dev/null && cd ../

View File

@@ -24,7 +24,7 @@ function build_docker_images() {
docker build --no-cache -t ${REGISTRY}/comps-base:${TAG} --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile .
popd && sleep 1s
git clone https://github.com/HabanaAI/vllm-fork.git && cd vllm-fork
VLLM_FORK_VER=v0.6.6.post1+Gaudi-1.20.0
VLLM_FORK_VER=v0.7.2+Gaudi-1.21.0
git checkout ${VLLM_FORK_VER} &> /dev/null && cd ../
echo "Build all the images with --no-cache, check docker_image_build.log for details..."

View File

@@ -24,7 +24,7 @@ function build_docker_images() {
docker build --no-cache -t ${REGISTRY}/comps-base:${TAG} --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile .
popd && sleep 1s
git clone https://github.com/HabanaAI/vllm-fork.git && cd vllm-fork
VLLM_FORK_VER=v0.6.6.post1+Gaudi-1.20.0
VLLM_FORK_VER=v0.7.2+Gaudi-1.21.0
git checkout ${VLLM_FORK_VER} &> /dev/null && cd ../
echo "Build all the images with --no-cache, check docker_image_build.log for details..."

View File

@@ -24,7 +24,7 @@ function build_docker_images() {
docker build --no-cache -t ${REGISTRY}/comps-base:${TAG} --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile .
popd && sleep 1s
git clone https://github.com/HabanaAI/vllm-fork.git && cd vllm-fork
VLLM_FORK_VER=v0.6.6.post1+Gaudi-1.20.0
VLLM_FORK_VER=v0.7.2+Gaudi-1.21.0
git checkout ${VLLM_FORK_VER} &> /dev/null && cd ../
echo "Build all the images with --no-cache, check docker_image_build.log for details..."

View File

@@ -24,7 +24,7 @@ function build_docker_images() {
docker build --no-cache -t ${REGISTRY}/comps-base:${TAG} --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile .
popd && sleep 1s
git clone https://github.com/HabanaAI/vllm-fork.git && cd vllm-fork
VLLM_FORK_VER=v0.6.6.post1+Gaudi-1.20.0
VLLM_FORK_VER=v0.7.2+Gaudi-1.21.0
git checkout ${VLLM_FORK_VER} &> /dev/null && cd ../
echo "Build all the images with --no-cache, check docker_image_build.log for details..."

View File

@@ -37,7 +37,7 @@ function build_docker_images() {
# Download Gaudi vllm of latest tag
git clone https://github.com/HabanaAI/vllm-fork.git && cd vllm-fork
VLLM_FORK_VER=v0.6.6.post1+Gaudi-1.20.0
VLLM_FORK_VER=v0.7.2+Gaudi-1.21.0
echo "Check out vLLM tag ${VLLM_FORK_VER}"
git checkout ${VLLM_FORK_VER} &> /dev/null && cd ../

View File

@@ -26,7 +26,7 @@ function build_docker_images() {
popd && sleep 1s
git clone https://github.com/HabanaAI/vllm-fork.git && cd vllm-fork
VLLM_FORK_VER=v0.6.6.post1+Gaudi-1.20.0
VLLM_FORK_VER=v0.7.2+Gaudi-1.21.0
git checkout ${VLLM_FORK_VER} &> /dev/null && cd ../
echo "Build all the images with --no-cache, check docker_image_build.log for details..."

View File

@@ -50,7 +50,7 @@ function build_docker_images() {
popd && sleep 1s
git clone https://github.com/HabanaAI/vllm-fork.git && cd vllm-fork
VLLM_FORK_VER=v0.6.6.post1+Gaudi-1.20.0
VLLM_FORK_VER=v0.7.2+Gaudi-1.21.0
git checkout ${VLLM_FORK_VER} &> /dev/null && cd ../
echo "Build all the images with --no-cache, check docker_image_build.log for details..."

View File

@@ -60,7 +60,7 @@ function build_vllm_docker_image() {
fi
cd ./vllm-fork
VLLM_FORK_VER=v0.6.6.post1+Gaudi-1.20.0
VLLM_FORK_VER=v0.7.2+Gaudi-1.21.0
git checkout ${VLLM_FORK_VER} &> /dev/null
docker build --no-cache -f Dockerfile.hpu -t $vllm_image --shm-size=128g . --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy
if [ $? -ne 0 ]; then

View File

@@ -27,7 +27,7 @@ function build_docker_images() {
popd && sleep 1s
git clone https://github.com/HabanaAI/vllm-fork.git && cd vllm-fork
VLLM_FORK_VER=v0.6.6.post1+Gaudi-1.20.0
VLLM_FORK_VER=v0.7.2+Gaudi-1.21.0
git checkout ${VLLM_FORK_VER} &> /dev/null && cd ../
service_list="visualqna visualqna-ui lvm nginx vllm-gaudi"