Docker Image #217
Closed
ArrichM
started this conversation in
Show and tell
Replies: 2 comments
-
Hello @ArrichM Thank in advance ! |
Beta Was this translation helpful? Give feedback.
0 replies
-
Done https://github.com/sgl-project/sglang?tab=readme-ov-file#method-3-using-docker |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Thank you so much for the great work you have done with sglang, it has so far been a really great experience for our use cases.
I created a Docker image to run an sglang server with flashinfer installed. The image is built on top of the official vllm image . To run it, use:
docker run --runtime nvidia --gpus all -v ~/.cache/huggingface:/root/.cache/huggingface -p 8000:8000 --ipc=host arrichm/sglang:latest --model-path mistralai/Mistral-7B-v0.1 --host 0.0.0.0
Here:
--model-path: specifies the model you want to serve
--host 0.0.0.0: Makes the endpoint accessible from outside the container.
If not already installed, you need to install the NVIDIA Container Toolkit before running.
Beta Was this translation helpful? Give feedback.
All reactions