Skip to content

[Bugfix] Qwen2.5-omni Qwen3-omni online gradio.py example fix#249

Merged
Gaohan123 merged 3 commits intovllm-project:mainfrom
david6666666:issue247
Dec 9, 2025
Merged

[Bugfix] Qwen2.5-omni Qwen3-omni online gradio.py example fix#249
Gaohan123 merged 3 commits intovllm-project:mainfrom
david6666666:issue247

Conversation

@david6666666
Copy link
Collaborator

PLEASE FILL IN THE PR DESCRIPTION HERE ENSURING ALL CHECKLIST ITEMS (AT THE BOTTOM) HAVE BEEN CONSIDERED.

Purpose

fix issue

Test Plan

python examples/online_serving/qwen2_5_omni/gradio_demo.py --model /workspace/models/Qwen/Qwen2.5-Omni-7B

Test Result

INFO:httpx:HTTP Request: GET http://localhost:7861/gradio_api/startup-events "HTTP/1.1 200 OK"

INFO:httpx:HTTP Request: HEAD http://localhost:7861/ "HTTP/1.1 200 OK"

* To create a public link, set `share=True` in `launch()`.

--------------------------------

[Stage-0] Received batch size=1, request_ids=0

--------------------------------

--------------------------------

[Stage-1] Received batch size=1, request_ids=0

--------------------------------

(EngineCore_DP0 pid=5102) /workspace/c00580271/cwq_branch/vllm-omni/vllm_omni/worker/gpu_model_runner.py:207: UserWarning: The given NumPy array is not writable, and PyTorch does not support non-writable tensors. This means writing to this tensor will result in undefined behavior. You may want to copy the array to protect its data or make it writable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at /pytorch/torch/csrc/utils/tensor_numpy.cpp:203.)

(EngineCore_DP0 pid=5102)   info_dict[k] = torch.from_numpy(arr)

--------------------------------

[Stage-2] Received batch size=1, request_ids=0

--------------------------------

(EngineCore_DP0 pid=5099) INFO:vllm_omni.model_executor.models.qwen2_5_omni.qwen2_5_omni:Currently, we do not use the chunked process, we only use the token2wav.process_chunk for the whole sequence. The stream mode will be implemented in the future.

INFO:vllm_omni.entrypoints.async_omni:[Summary] {'e2e_requests': 1, 'e2e_total_time_ms': 25743.138313293457, 'e2e_sum_time_ms': 25742.77091026306, 'e2e_total_tokens': 0, 'e2e_avg_time_per_request_ms': 25742.77091026306, 'e2e_avg_tokens_per_s': 0.0, 'wall_time_ms': 25743.138313293457, 'final_stage_id': 2, 'stages': [{'stage_id': 0, 'requests': 1, 'tokens': 49, 'total_time_ms': 1740.2911186218262, 'avg_time_per_request_ms': 1740.2911186218262, 'avg_tokens_per_s': 28.156208737538204}, {'stage_id': 1, 'requests': 1, 'tokens': 823, 'total_time_ms': 20198.933124542236, 'avg_time_per_request_ms': 20198.933124542236, 'avg_tokens_per_s': 40.744726215268926}, {'stage_id': 2, 'requests': 1, 'tokens': 0, 'total_time_ms': 3762.7434730529785, 'avg_time_per_request_ms': 3762.7434730529785, 'avg_tokens_per_s': 0.0}], 'transfers': [{'from_stage': 0, 'to_stage': 1, 'samples': 1, 'total_bytes': 1650152, 'total_time_ms': 1.3203620910644531, 'tx_mbps': 9998.178597628024, 'rx_samples': 1, 'rx_total_bytes': 1650152, 'rx_total_time_ms': 2.2895336151123047, 'rx_mbps': 5765.897435558055, 'total_samples': 1, 'total_transfer_time_ms': 4.13203239440918, 'total_mbps': 3194.8481376529917}, {'from_stage': 1, 'to_stage': 2, 'samples': 1, 'total_bytes': 2503, 'total_time_ms': 0.3561973571777344, 'tx_mbps': 56.216026302543504, 'rx_samples': 1, 'rx_total_bytes': 2503, 'rx_total_time_ms': 0.03147125244140625, 'rx_mbps': 636.2632067878787, 'total_samples': 1, 'total_transfer_time_ms': 1.3468265533447266, 'total_mbps': 14.867541741193131}]}

/workspace/c00580271/.venv/lib/python3.12/site-packages/gradio/processing_utils.py:688: UserWarning: Trying to convert audio automatically from float32 to 16-bit int format.

  warnings.warn(warning.format(data.dtype))
image
Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft.

BEFORE SUBMITTING, PLEASE READ https://github.com/vllm-project/vllm-omni/blob/main/CONTRIBUTING.md (anything written below this line will be removed by GitHub Actions)

Signed-off-by: David Chen <[email protected]>
Signed-off-by: David Chen <[email protected]>
@david6666666
Copy link
Collaborator Author

@Gaohan123 @SamitHuang PTAL,
CI failure was due to Diffusion Model Test Canceled.

Signed-off-by: David Chen <[email protected]>
Copy link
Collaborator

@Gaohan123 Gaohan123 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM. Thanks for the nice catch!

model=base_args.model,
stage_configs_path=getattr(base_args, "stage_configs_path", None),
log_stats=False,
log_file=None,
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Maybe we can add parser of arguments like PR#206 such that users can set them rather than hack statically here.

@Gaohan123 Gaohan123 merged commit c45cdea into vllm-project:main Dec 9, 2025
4 checks passed
@david6666666 david6666666 deleted the issue247 branch December 11, 2025 03:03
LawJarp-A pushed a commit to LawJarp-A/vllm-omni that referenced this pull request Dec 12, 2025
LawJarp-A pushed a commit to LawJarp-A/vllm-omni that referenced this pull request Dec 12, 2025
faaany pushed a commit to faaany/vllm-omni that referenced this pull request Dec 19, 2025
princepride pushed a commit to princepride/vllm-omni that referenced this pull request Jan 10, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

[Bug]: Qwen2.5-omni Qwen3-omni gradio.py, AsyncOmni.__init__() missing 1 required positional argument: 'cli_args'

2 participants