remove vllm hpu commit id limit (#1139)

Signed-off-by: Xinyao Wang <xinyao.wang@intel.com>
This commit is contained in:
XinyaoWa
2025-01-14 11:05:05 +08:00
committed by GitHub
parent 1cc4d2119d
commit afb8a28b2f
5 changed files with 0 additions and 5 deletions

View File

@@ -65,7 +65,6 @@ jobs:
fi
if [[ $(grep -c "vllm-gaudi:" ${docker_compose_yml}) != 0 ]]; then
git clone https://github.com/HabanaAI/vllm-fork.git vllm-fork
cd vllm-fork && git checkout 3c39626 && cd ../
fi
- name: Get build list
id: get-build-list

View File

@@ -37,7 +37,6 @@ fi
if [ "$hw_mode" = "hpu" ]; then
git clone https://github.com/HabanaAI/vllm-fork.git
cd ./vllm-fork/
git checkout 3c39626
docker build -f Dockerfile.hpu -t opea/vllm-gaudi:latest --shm-size=128g . --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy
cd ..
rm -rf vllm-fork

View File

@@ -12,7 +12,6 @@ function build_docker_images() {
cd $WORKPATH
git clone https://github.com/HabanaAI/vllm-fork.git
cd vllm-fork/
git checkout 3c39626
docker build --no-cache -f Dockerfile.hpu -t opea/vllm-gaudi:latest --shm-size=128g .
if [ $? -ne 0 ]; then
echo "opea/vllm-gaudi built fail"

View File

@@ -12,7 +12,6 @@ function build_docker_images() {
cd $WORKPATH
git clone https://github.com/HabanaAI/vllm-fork.git
cd vllm-fork/
git checkout 3c39626
docker build --no-cache -f Dockerfile.hpu -t opea/vllm-gaudi:latest --shm-size=128g .
if [ $? -ne 0 ]; then
echo "opea/vllm-gaudi built fail"

View File

@@ -12,7 +12,6 @@ function build_docker_images() {
cd $WORKPATH
git clone https://github.com/HabanaAI/vllm-fork.git
cd vllm-fork/
git checkout 3c39626
docker build --no-cache -f Dockerfile.hpu -t opea/vllm-gaudi:comps --shm-size=128g .
if [ $? -ne 0 ]; then
echo "opea/vllm-gaudi built fail"