File tree Expand file tree Collapse file tree 15 files changed +11
-64
lines changed
docker_compose/intel/hpu/gaudi
docker_compose/intel/hpu/gaudi
docker_compose/intel/hpu/gaudi
docker_compose/intel/hpu/gaudi
docker_compose/intel/hpu/gaudi Expand file tree Collapse file tree 15 files changed +11
-64
lines changed Original file line number Diff line number Diff line change @@ -128,7 +128,7 @@ services:
128128 OPENAI_API_BASE_URLS : ${SUPERVISOR_AGENT_ENDPOINT}
129129 ENABLE_OLLAMA_API : False
130130 vllm-service :
131- image : ${REGISTRY:- opea} /vllm-gaudi:${TAG:-latest}
131+ image : opea/vllm-gaudi:1.4
132132 container_name : vllm-gaudi-server
133133 ports :
134134 - " 8086:8000"
Original file line number Diff line number Diff line change @@ -17,12 +17,6 @@ services:
1717 dockerfile : ./docker/Dockerfile
1818 extends : agent
1919 image : ${REGISTRY:-opea}/agent-ui:${TAG:-latest}
20- vllm-gaudi :
21- build :
22- context : vllm-fork
23- dockerfile : Dockerfile.hpu
24- extends : agent
25- image : ${REGISTRY:-opea}/vllm-gaudi:${TAG:-latest}
2620 vllm-rocm :
2721 build :
2822 context : GenAIComps
Original file line number Diff line number Diff line change @@ -40,12 +40,8 @@ function build_agent_docker_image_gaudi_vllm() {
4040 cd $WORKDIR /GenAIExamples/AgentQnA/docker_image_build/
4141 get_genai_comps
4242
43- git clone https://github.com/HabanaAI/vllm-fork.git && cd vllm-fork
44- VLLM_FORK_VER=v0.6.6.post1+Gaudi-1.20.0
45- git checkout ${VLLM_FORK_VER} & > /dev/null && cd ../
46-
4743 echo " Build agent image with --no-cache..."
48- service_list=" agent agent-ui vllm-gaudi "
44+ service_list=" agent agent-ui"
4945 docker compose -f build.yaml build ${service_list} --no-cache
5046}
5147
@@ -83,6 +79,7 @@ function main() {
8379 " rocm_vllm" )
8480 echo " ==================== Build agent docker image for ROCm VLLM ===================="
8581 build_agent_docker_image_rocm_vllm
82+ docker image ls | grep vllm
8683 ;;
8784 " gaudi_vllm" )
8885 echo " ==================== Build agent docker image for Gaudi ===================="
@@ -97,8 +94,6 @@ function main() {
9794 exit 1
9895 ;;
9996 esac
100-
101- docker image ls | grep vllm
10297}
10398
10499main $1
Original file line number Diff line number Diff line change 33
44services :
55 vllm-service :
6- image : ${REGISTRY:- opea} /vllm-gaudi:${TAG:-latest}
6+ image : opea/vllm-gaudi:1.4
77 container_name : vllm-gaudi-server
88 ports :
99 - " 8028:80"
Original file line number Diff line number Diff line change @@ -78,12 +78,6 @@ services:
7878 dockerfile : comps/third_parties/vllm/src/Dockerfile.amd_gpu
7979 extends : codegen
8080 image : ${REGISTRY:-opea}/vllm-rocm:${TAG:-latest}
81- vllm-gaudi :
82- build :
83- context : vllm-fork
84- dockerfile : Dockerfile.hpu
85- extends : codegen
86- image : ${REGISTRY:-opea}/vllm-gaudi:${TAG:-latest}
8781 dataprep :
8882 build :
8983 context : GenAIComps
Original file line number Diff line number Diff line change @@ -25,14 +25,8 @@ function build_docker_images() {
2525 docker build --no-cache -t ${REGISTRY} /comps-base:${TAG} --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile .
2626 popd && sleep 1s
2727
28- # Download Gaudi vllm of latest tag
29- git clone https://github.com/HabanaAI/vllm-fork.git && cd vllm-fork
30- VLLM_FORK_VER=v0.6.6.post1+Gaudi-1.20.0
31- echo " Check out vLLM tag ${VLLM_FORK_VER} "
32- git checkout ${VLLM_FORK_VER} & > /dev/null && cd ../
33-
3428 echo " Build all the images with --no-cache, check docker_image_build.log for details..."
35- service_list=" codegen codegen-gradio-ui llm-textgen vllm-gaudi dataprep retriever embedding"
29+ service_list=" codegen codegen-gradio-ui llm-textgen dataprep retriever embedding"
3630 docker compose -f build.yaml build ${service_list} --no-cache > ${LOG_PATH} /docker_image_build.log
3731
3832 docker images && sleep 1s
Original file line number Diff line number Diff line change 33
44services :
55 vllm-service :
6- image : ${REGISTRY:- opea} /vllm-gaudi:${TAG:-latest}
6+ image : opea/vllm-gaudi:1.4
77 container_name : codetrans-gaudi-vllm-service
88 ports :
99 - " 8008:80"
Original file line number Diff line number Diff line change @@ -25,12 +25,6 @@ services:
2525 dockerfile : comps/llms/src/text-generation/Dockerfile
2626 extends : codetrans
2727 image : ${REGISTRY:-opea}/llm-textgen:${TAG:-latest}
28- vllm-gaudi :
29- build :
30- context : vllm-fork
31- dockerfile : Dockerfile.hpu
32- extends : codetrans
33- image : ${REGISTRY:-opea}/vllm-gaudi:${TAG:-latest}
3428 nginx :
3529 build :
3630 context : GenAIComps
Original file line number Diff line number Diff line change @@ -25,12 +25,8 @@ function build_docker_images() {
2525 docker build --no-cache -t ${REGISTRY} /comps-base:${TAG} --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile .
2626 popd && sleep 1s
2727
28- git clone https://github.com/HabanaAI/vllm-fork.git && cd vllm-fork
29- VLLM_FORK_VER=v0.6.6.post1+Gaudi-1.20.0
30- git checkout ${VLLM_FORK_VER} & > /dev/null && cd ../
31-
3228 echo " Build all the images with --no-cache, check docker_image_build.log for details..."
33- service_list=" codetrans codetrans-ui llm-textgen vllm-gaudi nginx"
29+ service_list=" codetrans codetrans-ui llm-textgen nginx"
3430 docker compose -f build.yaml build ${service_list} --no-cache > ${LOG_PATH} /docker_image_build.log
3531
3632 docker images && sleep 1s
Original file line number Diff line number Diff line change 33
44services :
55 vllm-service :
6- image : ${REGISTRY:- opea} /vllm-gaudi:${TAG:-latest}
6+ image : opea/vllm-gaudi:1.4
77 container_name : docsum-gaudi-vllm-service
88 ports :
99 - ${LLM_ENDPOINT_PORT:-8008}:80
You can’t perform that action at this time.
0 commit comments