Skip to content

[Bug]: vllm-omni(v0.12.0) results of talker model of qwen2.5-omni are incorrect when running with enforce eager being False #912

@Jinxinxiang5525

Description

@Jinxinxiang5525

Your current environment

The output of python collect_env.py
Your output of `python collect_env.py` here

Your code version

The commit id or version of vllm

The commit id or version of vllm-omni

🐛 Describe the bug

The tokens output by the talker model of qwen2.5-omni are incorrect when enforce eager (in qwen2_5_omni.yaml) is set to false (running compile on NPU), but it's correct for thinker model under the same parameters

Before submitting a new issue...

  • Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.

Metadata

Metadata

Assignees

Labels

bugSomething isn't working

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions