Skip to content

Commit b611e10

Browse files
authored
Use built image opea/vllm-gaudi:1.4 for AgentQnA, CodeGen, CodeTrans, DocSum and VisualQnA. (opea-project#2277)
Signed-off-by: ZePan110 <[email protected]>
1 parent b363635 commit b611e10

File tree

15 files changed

+11
-64
lines changed

15 files changed

+11
-64
lines changed

AgentQnA/docker_compose/intel/hpu/gaudi/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -128,7 +128,7 @@ services:
128128
OPENAI_API_BASE_URLS: ${SUPERVISOR_AGENT_ENDPOINT}
129129
ENABLE_OLLAMA_API: False
130130
vllm-service:
131-
image: ${REGISTRY:-opea}/vllm-gaudi:${TAG:-latest}
131+
image: opea/vllm-gaudi:1.4
132132
container_name: vllm-gaudi-server
133133
ports:
134134
- "8086:8000"

AgentQnA/docker_image_build/build.yaml

Lines changed: 0 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -17,12 +17,6 @@ services:
1717
dockerfile: ./docker/Dockerfile
1818
extends: agent
1919
image: ${REGISTRY:-opea}/agent-ui:${TAG:-latest}
20-
vllm-gaudi:
21-
build:
22-
context: vllm-fork
23-
dockerfile: Dockerfile.hpu
24-
extends: agent
25-
image: ${REGISTRY:-opea}/vllm-gaudi:${TAG:-latest}
2620
vllm-rocm:
2721
build:
2822
context: GenAIComps

AgentQnA/tests/step1_build_images.sh

Lines changed: 2 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -40,12 +40,8 @@ function build_agent_docker_image_gaudi_vllm() {
4040
cd $WORKDIR/GenAIExamples/AgentQnA/docker_image_build/
4141
get_genai_comps
4242

43-
git clone https://github.com/HabanaAI/vllm-fork.git && cd vllm-fork
44-
VLLM_FORK_VER=v0.6.6.post1+Gaudi-1.20.0
45-
git checkout ${VLLM_FORK_VER} &> /dev/null && cd ../
46-
4743
echo "Build agent image with --no-cache..."
48-
service_list="agent agent-ui vllm-gaudi"
44+
service_list="agent agent-ui"
4945
docker compose -f build.yaml build ${service_list} --no-cache
5046
}
5147

@@ -83,6 +79,7 @@ function main() {
8379
"rocm_vllm")
8480
echo "==================== Build agent docker image for ROCm VLLM ===================="
8581
build_agent_docker_image_rocm_vllm
82+
docker image ls | grep vllm
8683
;;
8784
"gaudi_vllm")
8885
echo "==================== Build agent docker image for Gaudi ===================="
@@ -97,8 +94,6 @@ function main() {
9794
exit 1
9895
;;
9996
esac
100-
101-
docker image ls | grep vllm
10297
}
10398

10499
main $1

CodeGen/docker_compose/intel/hpu/gaudi/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33

44
services:
55
vllm-service:
6-
image: ${REGISTRY:-opea}/vllm-gaudi:${TAG:-latest}
6+
image: opea/vllm-gaudi:1.4
77
container_name: vllm-gaudi-server
88
ports:
99
- "8028:80"

CodeGen/docker_image_build/build.yaml

Lines changed: 0 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -78,12 +78,6 @@ services:
7878
dockerfile: comps/third_parties/vllm/src/Dockerfile.amd_gpu
7979
extends: codegen
8080
image: ${REGISTRY:-opea}/vllm-rocm:${TAG:-latest}
81-
vllm-gaudi:
82-
build:
83-
context: vllm-fork
84-
dockerfile: Dockerfile.hpu
85-
extends: codegen
86-
image: ${REGISTRY:-opea}/vllm-gaudi:${TAG:-latest}
8781
dataprep:
8882
build:
8983
context: GenAIComps

CodeGen/tests/test_compose_on_gaudi.sh

Lines changed: 1 addition & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -25,14 +25,8 @@ function build_docker_images() {
2525
docker build --no-cache -t ${REGISTRY}/comps-base:${TAG} --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile .
2626
popd && sleep 1s
2727

28-
# Download Gaudi vllm of latest tag
29-
git clone https://github.com/HabanaAI/vllm-fork.git && cd vllm-fork
30-
VLLM_FORK_VER=v0.6.6.post1+Gaudi-1.20.0
31-
echo "Check out vLLM tag ${VLLM_FORK_VER}"
32-
git checkout ${VLLM_FORK_VER} &> /dev/null && cd ../
33-
3428
echo "Build all the images with --no-cache, check docker_image_build.log for details..."
35-
service_list="codegen codegen-gradio-ui llm-textgen vllm-gaudi dataprep retriever embedding"
29+
service_list="codegen codegen-gradio-ui llm-textgen dataprep retriever embedding"
3630
docker compose -f build.yaml build ${service_list} --no-cache > ${LOG_PATH}/docker_image_build.log
3731

3832
docker images && sleep 1s

CodeTrans/docker_compose/intel/hpu/gaudi/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33

44
services:
55
vllm-service:
6-
image: ${REGISTRY:-opea}/vllm-gaudi:${TAG:-latest}
6+
image: opea/vllm-gaudi:1.4
77
container_name: codetrans-gaudi-vllm-service
88
ports:
99
- "8008:80"

CodeTrans/docker_image_build/build.yaml

Lines changed: 0 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -25,12 +25,6 @@ services:
2525
dockerfile: comps/llms/src/text-generation/Dockerfile
2626
extends: codetrans
2727
image: ${REGISTRY:-opea}/llm-textgen:${TAG:-latest}
28-
vllm-gaudi:
29-
build:
30-
context: vllm-fork
31-
dockerfile: Dockerfile.hpu
32-
extends: codetrans
33-
image: ${REGISTRY:-opea}/vllm-gaudi:${TAG:-latest}
3428
nginx:
3529
build:
3630
context: GenAIComps

CodeTrans/tests/test_compose_on_gaudi.sh

Lines changed: 1 addition & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -25,12 +25,8 @@ function build_docker_images() {
2525
docker build --no-cache -t ${REGISTRY}/comps-base:${TAG} --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile .
2626
popd && sleep 1s
2727

28-
git clone https://github.com/HabanaAI/vllm-fork.git && cd vllm-fork
29-
VLLM_FORK_VER=v0.6.6.post1+Gaudi-1.20.0
30-
git checkout ${VLLM_FORK_VER} &> /dev/null && cd ../
31-
3228
echo "Build all the images with --no-cache, check docker_image_build.log for details..."
33-
service_list="codetrans codetrans-ui llm-textgen vllm-gaudi nginx"
29+
service_list="codetrans codetrans-ui llm-textgen nginx"
3430
docker compose -f build.yaml build ${service_list} --no-cache > ${LOG_PATH}/docker_image_build.log
3531

3632
docker images && sleep 1s

DocSum/docker_compose/intel/hpu/gaudi/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33

44
services:
55
vllm-service:
6-
image: ${REGISTRY:-opea}/vllm-gaudi:${TAG:-latest}
6+
image: opea/vllm-gaudi:1.4
77
container_name: docsum-gaudi-vllm-service
88
ports:
99
- ${LLM_ENDPOINT_PORT:-8008}:80

0 commit comments

Comments
 (0)