Skip to content

[feat]: adapt batch request for flux#1028

Merged
david6666666 merged 1 commit intovllm-project:mainfrom
nuclearwu:request
Jan 29, 2026
Merged

[feat]: adapt batch request for flux#1028
david6666666 merged 1 commit intovllm-project:mainfrom
nuclearwu:request

Conversation

@nuclearwu
Copy link
Contributor

@nuclearwu nuclearwu commented Jan 28, 2026

Signed-off-by: wuzhongjian [email protected]

PLEASE FILL IN THE PR DESCRIPTION HERE ENSURING ALL CHECKLIST ITEMS (AT THE BOTTOM) HAVE BEEN CONSIDERED.

Purpose

adapt batch request for flux #853 and #797

Test Plan

Vllm-Omni:

  1. offline_inference
python examples/offline_inference/text_to_image/text_to_image.py \
  --model black-forest-labs/FLUX.1-dev \
  --prompt "Beautiful illustration of The ocean. in a serene landscape, magic realism, narrative realism, beautiful matte painting, heavenly lighting, retrowave, 4 k hd wallpaper" \
  --seed 42 \
  --cfg_scale 4.0 \
  --tensor_parallel_size 4 \
  --num_images_per_prompt 1 \
  --num_inference_steps 50 \
  --guidance_scale 4.0 \
  --height 1024\
  --width 1024\
  --output outputs/ocean.png
  1. online_inference
export VLLM_WORKER_MULTIPROC_METHOD=spawn
MODEL_NAME_OR_PATH=/workspace/cache/ymttest/johnjan/models/black-forest-labs/FLUX___1-dev/

vllm serve ${MODEL_NAME_OR_PATH} \
   --omni \
   --port 8092 \
   --vae_use_slicing \
   --vae_use_tiling
curl -s http://localhost:8092/v1/chat/completions \
  -H "Content-Type: application/json" \
  -d '{
    "messages": [
      {"role": "user", "content": "Beautiful illustration of The ocean. in a serene landscape, magic realism, narrative realism, beautiful matte painting, heavenly lighting, retrowave, 4 k hd wallpaper"}
    ],
    "extra_body": {
      "height": 1024,
      "width": 1024,
      "num_inference_steps": 50,
      "guidance_scale": 4.0,
      "seed": 42
    }
  }' | jq -r '.choices[0].message.content[0].image_url.url' | cut -d',' -f2- | base64 -d > ocean-serve.png
  1. Benchmark
python3 benchmarks/diffusion/diffusion_benchmark_serving.py \
	--base-url http://localhost:8092 \
	--model /workspace/cache/ymttest/johnjan/models/black-forest-labs/FLUX___1-dev/ \
	--task t2i \
	--dataset vbench \
	--num-prompts 100

Test Result

1.offline_inference
59427de38415eaca375f4923a8f7d18b

2.online_inference
59427de38415eaca375f4923a8f7d18b

  1. Benchmark
    daf7d7d4d0903d87ce7e83680e42b5b8

Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft.

BEFORE SUBMITTING, PLEASE READ https://github.com/vllm-project/vllm-omni/blob/main/CONTRIBUTING.md (anything written below this line will be removed by GitHub Actions)

@nuclearwu nuclearwu mentioned this pull request Jan 28, 2026
5 tasks
@nuclearwu
Copy link
Contributor Author

@david6666666
Copy link
Collaborator

@fhfuih ptal thx

Copy link

@chatgpt-codex-connector chatgpt-codex-connector bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

💡 Codex Review

Here are some automated review suggestions for this pull request.

Reviewed commit: f108c48697

ℹ️ About Codex in GitHub

Codex has been enabled to automatically review pull requests in this repo. Reviews are triggered when you

  • Open a pull request for review
  • Mark a draft as ready
  • Comment "@codex review".

If Codex has suggestions, it will comment; otherwise it will react with 👍.

When you sign up for Codex through ChatGPT, Codex can also answer questions or update the PR, like "@codex address that feedback".

@hsliuustc0106
Copy link
Collaborator

add test plan&results please

@ZJY0516 ZJY0516 added this to the v0.14.0 milestone Jan 28, 2026
@fhfuih
Copy link
Contributor

fhfuih commented Jan 29, 2026

Looks good to me. Thanks for keeping up with the recent code update

@nuclearwu
Copy link
Contributor Author

nuclearwu commented Jan 29, 2026

add test plan&results please

@hsliuustc0106 done, please review, thanks

@nuclearwu
Copy link
Contributor Author

cc @david6666666 @ZJY0516

@hsliuustc0106
Copy link
Collaborator

@hsliuustc0106 hsliuustc0106 added the ready label to trigger buildkite CI label Jan 29, 2026
@nuclearwu
Copy link
Contributor Author

could you provide benchmark results using https://github.com/vllm-project/vllm-omni/tree/main/benchmarks/diffusion

@hsliuustc0106 done, ptal thx

@david6666666
Copy link
Collaborator

LGTM, I've already tested it locally, thanks for the fix.

@david6666666 david6666666 merged commit ee0f7a7 into vllm-project:main Jan 29, 2026
7 checks passed
dongbo910220 pushed a commit to dongbo910220/vllm-omni that referenced this pull request Feb 1, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ready label to trigger buildkite CI

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants