Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
58 commits
Select commit Hold shift + click to select a range
7a69c19
feat/UltraFlux-v1-image
erfgss Jan 13, 2026
0818688
feat/UltraFlux-v1-image
erfgss Jan 13, 2026
d117835
feat/UltraFlux-v1-image
erfgss Jan 13, 2026
159fba2
[Bugfix] Raise ValueError when joint_strategy='rear' and causal=True …
mxuax Jan 13, 2026
f130723
[Feat] add vllm-omni version collection (#740)
sihyeonn Jan 13, 2026
060bb48
feat/UltraFlux-v1-image
erfgss Jan 13, 2026
cc50071
feat/UltraFlux-v1-image
erfgss Jan 13, 2026
793ad0d
feat/UltraFlux-v1-image
erfgss Jan 13, 2026
098473c
feat/UltraFlux-v1-image
erfgss Jan 13, 2026
789930c
feat/UltraFlux-v1-image
erfgss Jan 13, 2026
059ec13
feat/UltraFlux-v1-image
erfgss Jan 13, 2026
3adc8db
feat/UltraFlux-v1-image
erfgss Jan 13, 2026
bb0618f
feat/UltraFlux-v1-image
erfgss Jan 13, 2026
074cb93
[Doc] refactor diffusion doc (#753)
ZJY0516 Jan 13, 2026
cc733a1
feat/UltraFlux-v1-image
erfgss Jan 13, 2026
31c245f
[Bugfix] Fix stable diffusion3 compatibility error (#772)
iwzbi Jan 13, 2026
2bab2c0
[Feature] Support Qwen3 Omni talker mtp batch inference (#722)
ZeldaHuang Jan 14, 2026
35b5a7a
[BugFix]Remove duplicate error handling for request results (#781)
liuyuhanalex Jan 14, 2026
28a5793
[CI] Add pytest markers in config files. (#719)
congw729 Jan 14, 2026
63febde
[Doc] Fix mkdocs. (#785)
congw729 Jan 15, 2026
5213905
[Bugfix] Fix generation artifacts of Qwen-Image-Edit-2511 and update …
SamitHuang Jan 15, 2026
91de0c4
[bugfix] Fix Wan2.2 I2V warmup failure by adding support_image_input …
linyueqian Jan 15, 2026
00b2e6c
[Misc] add wechat group and star history on README (#801)
david6666666 Jan 15, 2026
473643b
[BugFix] Fix incorrect mrope positions under cuda graph (#803)
ZeldaHuang Jan 15, 2026
ae12cc8
[BugFix] Qwen2.5-omni supress end token and won't stop (#773)
yinpeiqi Jan 15, 2026
cc0e014
[Feature] Flash Attention to Support Attention Mask (#760)
wtomin Jan 16, 2026
99f3c80
feat/UltraFlux-v1-image
erfgss Jan 19, 2026
848ac69
feat/UltraFlux-v1-image
erfgss Jan 19, 2026
4dfc48d
feat/UltraFlux-v1-image
erfgss Jan 19, 2026
54b57f8
[bugfix] use unipc scheduler for Wan 2.2 (#804)
linyueqian Jan 16, 2026
2be4577
[Test] Add full test for Qwen3-Omni-30B-A3B-Instruct (#720)
yenuo26 Jan 16, 2026
102d114
[Bagel] Support Cache-Dit (#736)
princepride Jan 17, 2026
ed9fb0a
[Perf] Optimize the Qwen2.5-Omni Model thinker-to-talker-proj with nn…
kechengliu97 Jan 17, 2026
2c43fe8
[Core]Add GPU Diffusion Runner (#822)
princepride Jan 17, 2026
bf5d081
feat/UltraFlux-v1-image
erfgss Jan 19, 2026
5042c97
feat/UltraFlux-v1-image
erfgss Jan 19, 2026
e45994c
[Model] add flux2 klein (#809)
erfgss Jan 19, 2026
b04bf98
feat/UltraFlux-v1-image
erfgss Jan 19, 2026
67a48fb
feat/UltraFlux-v1-image
erfgss Jan 19, 2026
47fd0c6
Merge branch 'main' into feat/UltraFlux-v1-image
erfgss Jan 19, 2026
3cc5cfb
feat/UltraFlux-v1-image
erfgss Jan 19, 2026
21e2132
feat/UltraFlux-v1-image
erfgss Jan 19, 2026
f626c02
feat/UltraFlux-v1-image
erfgss Jan 19, 2026
8ee88ae
feat/UltraFlux-v1-image
erfgss Jan 19, 2026
efacd5e
Merge branch 'main' into feat/UltraFlux-v1-image
erfgss Jan 19, 2026
c0c5ace
Merge branch 'main' into feat/UltraFlux-v1-image
erfgss Jan 19, 2026
c088962
feat/UltraFlux-v1-image
erfgss Jan 19, 2026
32fe65d
feat/UltraFlux-v1-image
erfgss Jan 19, 2026
f085fe3
feat/UltraFlux-v1-image
erfgss Jan 19, 2026
e643b8b
feat/UltraFlux-v1-image
erfgss Jan 19, 2026
781a16a
feat/UltraFlux-v1-image
erfgss Jan 19, 2026
f58a207
feat/UltraFlux-v1-image
erfgss Jan 19, 2026
43eab3a
feat/UltraFlux-v1-image
erfgss Jan 19, 2026
d8a81e5
Merge branch 'main' into feat/UltraFlux-v1-image
erfgss Jan 22, 2026
7346450
Merge branch 'main' into feat/UltraFlux-v1-image
erfgss Jan 23, 2026
7ac98e2
Merge branch 'main' into feat/UltraFlux-v1-image
erfgss Feb 3, 2026
d24ca44
Merge branch 'main' into feat/UltraFlux-v1-image
erfgss Feb 5, 2026
3d8dfb3
Merge branch 'main' into feat/UltraFlux-v1-image
erfgss Feb 10, 2026
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
5 changes: 1 addition & 4 deletions docs/models/supported_models.md
Original file line number Diff line number Diff line change
Expand Up @@ -37,10 +37,7 @@ th {
|`Flux2KleinPipeline` | FLUX.2-klein | `black-forest-labs/FLUX.2-klein-4B`, `black-forest-labs/FLUX.2-klein-9B` |
|`FluxPipeline` | FLUX.1-dev | `black-forest-labs/FLUX.1-dev` |
|`StableAudioPipeline` | Stable-Audio-Open | `stabilityai/stable-audio-open-1.0` |
|`Qwen3TTSForConditionalGeneration` | Qwen3-TTS-12Hz-1.7B-CustomVoice | `Qwen/Qwen3-TTS-12Hz-1.7B-CustomVoice` |
|`Qwen3TTSForConditionalGeneration` | Qwen3-TTS-12Hz-1.7B-VoiceDesign | `Qwen/Qwen3-TTS-12Hz-1.7B-VoiceDesign` |
|`Qwen3TTSForConditionalGeneration` | Qwen3-TTS-12Hz-1.7B-Base | `Qwen/Qwen3-TTS-12Hz-0.6B-Base` |

|`UltraFluxPipeline` | UltraFlux-v1 | `Owen777/UltraFlux-v1` |
Copy link
Contributor

@lishunyang12 lishunyang12 Feb 22, 2026

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks like this diff might have accidentally removed the three Qwen3-TTS entries — probably a rebase artifact? The UltraFlux line should be added alongside them.


## List of Supported Models for NPU

Expand Down
55 changes: 53 additions & 2 deletions vllm_omni/diffusion/cache/cache_dit_backend.py
Original file line number Diff line number Diff line change
Expand Up @@ -346,6 +346,55 @@ def refresh_cache_context(pipeline: Any, num_inference_steps: int, verbose: bool
return refresh_cache_context


def enable_cache_for_ultraflux(pipeline: Any, cache_config: Any) -> Callable[[int], None]:
db_cache_config = _build_db_cache_config(cache_config)

calibrator = None
if cache_config.enable_taylorseer:
taylorseer_order = cache_config.taylorseer_order
calibrator = TaylorSeerCalibratorConfig(taylorseer_order=taylorseer_order)
logger.info(f"TaylorSeer enabled with order={taylorseer_order}")

modifier = ParamsModifier(cache_config=db_cache_config, calibrator_config=calibrator)

logger.info(
"Enabling cache-dit on Flux transformer blocks with BlockAdapter: "
f"Fn={db_cache_config.Fn_compute_blocks}, "
f"Bn={db_cache_config.Bn_compute_blocks}, "
f"W={db_cache_config.max_warmup_steps}, "
)

transformer = pipeline.transformer
cache_dit.enable_cache(
BlockAdapter(
transformer=transformer,
blocks=[transformer.transformer_blocks, transformer.single_transformer_blocks],
forward_pattern=[ForwardPattern.Pattern_1, ForwardPattern.Pattern_1],
params_modifiers=[modifier],
),
cache_config=db_cache_config,
)

def refresh_cache_context(pipeline: Any, num_inference_steps: int, verbose: bool = True) -> None:
if cache_config.scm_steps_mask_policy is None:
cache_dit.refresh_context(pipeline.transformer, num_inference_steps=num_inference_steps, verbose=verbose)
else:
cache_dit.refresh_context(
pipeline.transformer,
cache_config=DBCacheConfig().reset(
num_inference_steps=num_inference_steps,
steps_computation_mask=cache_dit.steps_mask(
mask_policy=cache_config.scm_steps_mask_policy,
total_steps=num_inference_steps,
),
steps_computation_policy=cache_config.scm_steps_policy,
),
verbose=verbose,
)

return refresh_cache_context


def enable_cache_for_sd3(pipeline: Any, cache_config: Any) -> Callable[[int], None]:
"""Enable cache-dit for StableDiffusion3Pipeline.

Expand Down Expand Up @@ -859,8 +908,9 @@ def refresh_cache_context(pipeline: Any, num_inference_steps: int, verbose: bool
"Wan22I2VPipeline": enable_cache_for_wan22,
"Wan22TI2VPipeline": enable_cache_for_wan22,
"FluxPipeline": enable_cache_for_flux,
"LongCatImagePipeline": enable_cache_for_longcat_image,
"LongCatImageEditPipeline": enable_cache_for_longcat_image,
"UltraFluxPipeline": enable_cache_for_ultraflux,
"LongcatImagePipeline": enable_cache_for_longcat_image,
Copy link
Contributor

@lishunyang12 lishunyang12 Feb 22, 2026

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The LongCatImagePipeline -> LongcatImagePipeline rename seems like a separate change. Might be worth mentioning in the PR description, or splitting it out if you prefer?

"LongcatImageEditPipeline": enable_cache_for_longcat_image,
"StableDiffusion3Pipeline": enable_cache_for_sd3,
"BagelPipeline": enable_cache_for_bagel,
}
Expand Down Expand Up @@ -978,4 +1028,5 @@ def may_enable_cache_dit(pipeline: Any, od_config: OmniDiffusionConfig) -> Optio

backend = CacheDiTBackend(od_config.cache_config)
backend.enable(pipeline)

return backend if backend.is_enabled() else None
8 changes: 8 additions & 0 deletions vllm_omni/diffusion/models/ultraflux-v1_image/__init__.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,8 @@
# SPDX-License-Identifier: Apache-2.0
# UltraFlux diffusion model integration for vLLM-Omni

from .pipeline_ultraflux import UltraFluxPipeline

__all__ = [
"UltraFluxPipeline",
]
Loading