Skip to content

[Refractor] Dependency refractored to vLLM v0.11.0#48

Merged
hsliuustc0106 merged 7 commits intovllm-project:mainfrom
Gaohan123:refractor_v2
Nov 7, 2025
Merged

[Refractor] Dependency refractored to vLLM v0.11.0#48
hsliuustc0106 merged 7 commits intovllm-project:mainfrom
Gaohan123:refractor_v2

Conversation

@Gaohan123
Copy link
Collaborator

PLEASE FILL IN THE PR DESCRIPTION HERE ENSURING ALL CHECKLIST ITEMS (AT THE BOTTOM) HAVE BEEN CONSIDERED.

Purpose

This PR refractors the dependency of vLLM to stable release version v0.11.0. It tries to resolve issue #39 and part of issue #33 .

Test Plan

Follow README.md in root directory.

Test Result

Adding requests: 100%|██████████████████████████████████████████████████████████████████| 1/1 [00:00<00:00, 34.36it/s]
Processed prompts: 100%|██████████| 1/1 [00:00<00:00,  1.01it/s, est. speed input: 67.59 toks/s, output: 49.43 toks/s]
Adding requests: 100%|█████████████████████████████████████████████████████████████████| 1/1 [00:00<00:00, 316.91it/s]
Processed prompts:   0%|                    | 0/1 [00:00<?, ?it/s, est. speed input: 0.00 toks/s, output: 0.00 toks/s](EngineCore_DP0 pid=27465) /root/gh/vllm_open_release/vllm/vllm_omni/worker/gpu_model_runner.py:153: UserWarning: The given NumPy array is not writable, and PyTorch does not support non-writable tensors. This means writing to this tensor will result in undefined behavior. You may want to copy the array to protect its data or make it writable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at /pytorch/torch/csrc/utils/tensor_numpy.cpp:203.)
(EngineCore_DP0 pid=27465)   info_dict[k] = torch.from_numpy(arr)
Processed prompts: 100%|███████████| 1/1 [00:12<00:00, 12.11s/it, est. speed input: 5.70 toks/s, output: 68.15 toks/s]
Adding requests: 100%|████████████████████████████████████████████████████████████████| 1/1 [00:00<00:00, 1736.77it/s]
Processed prompts: 100%|██████████| 1/1 [00:02<00:00,  2.03s/it, est. speed input: 406.90 toks/s, output: 0.00 toks/s]
Request ID: 0, Text Output: Well, it usually has input modules for data, processing units like neural networks or algorithms, output for generated audio, and scalability through parallel computing or distributed systems.If you want to know more about any part of this, feel free to ask.
Request ID: 0, Saved audio to output_audio/output_0.wav

Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft.

BEFORE SUBMITTING, PLEASE READ https://github.com/hsliuustc0106/vllm-omni/blob/main/CONTRIBUTING.md (anything written below this line will be removed by GitHub Actions)

@Gaohan123 Gaohan123 requested a review from tzhouam November 4, 2025 14:15
Signed-off-by: Gaohan123 <[email protected]>
@Gaohan123
Copy link
Collaborator Author

@hsliuustc0106

Signed-off-by: Gaohan123 <[email protected]>
Signed-off-by: Gaohan123 <[email protected]>
@hsliuustc0106
Copy link
Collaborator

lgtm
approve

@hsliuustc0106 hsliuustc0106 merged commit 86841b6 into vllm-project:main Nov 7, 2025
1 check passed
@Gaohan123 Gaohan123 deleted the refractor_v2 branch November 18, 2025 13:13
princepride pushed a commit to princepride/vllm-omni that referenced this pull request Jan 10, 2026
[Refractor] Dependency refractored to vLLM v0.11.0
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants