[feat]: adapt batch request for flux#1028
Conversation
Signed-off-by: wuzhongjian <[email protected]>
|
@fhfuih ptal thx |
There was a problem hiding this comment.
💡 Codex Review
Here are some automated review suggestions for this pull request.
Reviewed commit: f108c48697
ℹ️ About Codex in GitHub
Codex has been enabled to automatically review pull requests in this repo. Reviews are triggered when you
- Open a pull request for review
- Mark a draft as ready
- Comment "@codex review".
If Codex has suggestions, it will comment; otherwise it will react with 👍.
When you sign up for Codex through ChatGPT, Codex can also answer questions or update the PR, like "@codex address that feedback".
|
add test plan&results please |
|
Looks good to me. Thanks for keeping up with the recent code update |
@hsliuustc0106 done, please review, thanks |
|
could you provide benchmark results using https://github.com/vllm-project/vllm-omni/tree/main/benchmarks/diffusion |
@hsliuustc0106 done, ptal thx |
|
LGTM, I've already tested it locally, thanks for the fix. |
Signed-off-by: wuzhongjian [email protected]
Signed-off-by: wuzhongjian [email protected]
PLEASE FILL IN THE PR DESCRIPTION HERE ENSURING ALL CHECKLIST ITEMS (AT THE BOTTOM) HAVE BEEN CONSIDERED.
Purpose
adapt batch request for flux #853 and #797
Test Plan
Vllm-Omni:
Test Result
1.offline_inference

2.online_inference

Essential Elements of an Effective PR Description Checklist
supported_models.mdandexamplesfor a new model.BEFORE SUBMITTING, PLEASE READ https://github.com/vllm-project/vllm-omni/blob/main/CONTRIBUTING.md (anything written below this line will be removed by GitHub Actions)