[Doc] Add guides for custom docker image build on NVIDIA CUDA [Skip-CI]#1386
[Doc] Add guides for custom docker image build on NVIDIA CUDA [Skip-CI]#1386loveysuby wants to merge 7 commits intovllm-project:mainfrom
Conversation
Signed-off-by: Hyoseop Song <crad_on25@naver.com>
| You can use this docker image to serve models the same way you would with in vLLM! To do so, make sure you overwrite the default entrypoint (`vllm serve --omni`) which works only for models supported in the vLLM-Omni project. | ||
|
|
||
| # --8<-- [end:pre-built-images] | ||
|
|
||
| # --8<-- [start:build-docker] | ||
|
|
||
| #### Build docker image | ||
|
|
||
| ```bash | ||
| DOCKER_BUILDKIT=1 docker build -f docker/Dockerfile.ci -t vllm-omni-cuda . |
There was a problem hiding this comment.
@congw729 Hi, I've written a guide for NVIDIA GPU users, but using the Dockerfile.ci as-is doesn't seem suitable for the purpose.
I have already verified the installation logic on an NVIDIA A100. Should I create a new, dedicated Dockerfile for users and re-test it? Let me know your thoughts, and I'll update the PR accordingly.
Switching to Draft for now.
There was a problem hiding this comment.
@congw729 Hi, I've written a guide for NVIDIA GPU users, but using the
Dockerfile.cias-is doesn't seem suitable for the purpose.I have already verified the installation logic on an NVIDIA A100. Should I create a new, dedicated Dockerfile for users and re-test it? Let me know your thoughts, and I'll update the PR accordingly.
Switching to Draft for now.
I think it's better to using a different Dockerfile. Docker.ci will install unnecessary packages for users.
There was a problem hiding this comment.
Yes, the Dockerfile.ci install the vllm-omni in dev mode, which will include some unnecessary packages.
|
|
||
| ```bash | ||
| docker run --runtime nvidia --gpus all \ | ||
| -v ~/.cache/huggingface:/root/.cache/huggingface \ |
There was a problem hiding this comment.
This model needs significant GPU memory ("verified on 2 x H100s" above). Worth noting that or using --gpus 2 in the example.
|
@vllm-omni-reviewer |
🤖 VLLM-Omni PR ReviewCode Review: Add guides for custom docker image build on NVIDIA CUDA1. OverviewThis PR adds documentation for building custom Docker images on NVIDIA CUDA, mirroring the existing AMD ROCm guide structure. The changes include:
Overall Assessment: Positive - The PR follows the existing documentation patterns and provides useful guidance for users who need custom Docker builds. 2. Code QualityStrengths
Minor Issues
3. Architecture & Design
4. Security & Safety
5. Testing & DocumentationTest Plan
Documentation Completeness
SuggestionConsider adding a brief note about what modifications users might want to make when building custom images (e.g., "Modify the source code before building to include custom changes"). 6. Specific Suggestions
|
|
@lishunyang12 @congw729 Thanks for the review feedback. I've created Once #1439 is merged, I'll update this documentation PR:
|
|
@vllm-omni-reviewer |
|
Hello, any updates? Currently v0.16.0 has already released |
|
@Gaohan123 I sent you a message on the vLLM Slack about this updates and #1439. Please take a look (cc: @tzhouam) |
…ker-build-on-nvidia-cuda
…project#1439) Signed-off-by: Hyoseop Song <crad_on25@naver.com> Signed-off-by: Hyoseop Song <crad_on25@naver.com>
|
@Gaohan123 @lishunyang12 PTAL: There was an image build test in the PR body, but since this had already been verified in #1439, I removed it. Please let me know if you have any requests for changes to the document content. |
| You can use this docker image to serve models the same way you would with in vLLM! To do so, make sure you overwrite the default entrypoint (`vllm serve --omni`) which works only for models supported in the vLLM-Omni project. | ||
|
|
||
| # --8<-- [end:pre-built-images] | ||
|
|
||
| # --8<-- [start:build-docker] | ||
|
|
||
| #### Build docker image | ||
|
|
||
| ```bash | ||
| DOCKER_BUILDKIT=1 docker build -f docker/Dockerfile.ci -t vllm-omni-cuda . |
There was a problem hiding this comment.
Yes, the Dockerfile.ci install the vllm-omni in dev mode, which will include some unnecessary packages.
| ```bash | ||
| DOCKER_BUILDKIT=1 docker build \ | ||
| -f docker/Dockerfile.cuda \ | ||
| --build-arg BASE_IMAGE=vllm/vllm-openai:v0.18.0 \ |
Signed-off-by: Hyoseop Song <crad_on25@naver.com> Signed-off-by: Hyoseop Song <crad_on25@naver.com>
PLEASE FILL IN THE PR DESCRIPTION HERE ENSURING ALL CHECKLIST ITEMS (AT THE BOTTOM) HAVE BEEN CONSIDERED.
Purpose
Added
NVIDIA CUDAbuild instructions to match the existing AMD ROCm guide.Documents how to use
docker/Dockerfile.cudafor custom builds, enabling source modifications andBASE_IMAGEcustomization. (added in #1439)Test Plan
Runtime Environment: NVIDIA A100-SXM4-80GB (CUDA 13.0 / Driver 580.82.07)
docker build --check -f docker/Dockerfile.cudawith differentBASE_IMAGEto specify vLLM base image.DOCKER_BUILDKIT=1 docker build \ --check \ -f docker/Dockerfile.cuda \ --build-arg BASE_IMAGE=vllm/vllm-openai:v0.18.0 \ -t vllm-omni-cuda .Test Result
Essential Elements of an Effective PR Description Checklist
supported_models.mdandexamplesfor a new model. Please runmkdocs serveto sync the documentation editions to./docs.BEFORE SUBMITTING, PLEASE READ https://github.com/vllm-project/vllm-omni/blob/main/CONTRIBUTING.md (anything written below this line will be removed by GitHub Actions)