-
-
Notifications
You must be signed in to change notification settings - Fork 11.7k
[Frontend][torch.compile] CompilationConfig Overhaul (#20283): name change compilation level to compilation mode, deprecation compilation level #26355
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Merged
ProExpertProg
merged 31 commits into
vllm-project:main
from
morrison-turnansky:issue-20283-level
Oct 15, 2025
Merged
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Contributor
|
This pull request has merge conflicts that must be resolved before it can be |
4753011 to
e347d4d
Compare
Contributor
Author
|
Issue created: #26911 |
1 task
bbartels
pushed a commit
to bbartels/vllm
that referenced
this pull request
Oct 16, 2025
): name change compilation level to compilation mode, deprecation compilation level (vllm-project#26355) Signed-off-by: morrison-turnansky <[email protected]> Signed-off-by: Morrison Turnansky <[email protected]> Co-authored-by: Luka Govedič <[email protected]> Signed-off-by: bbartels <[email protected]>
lywa1998
pushed a commit
to lywa1998/vllm
that referenced
this pull request
Oct 20, 2025
): name change compilation level to compilation mode, deprecation compilation level (vllm-project#26355) Signed-off-by: morrison-turnansky <[email protected]> Signed-off-by: Morrison Turnansky <[email protected]> Co-authored-by: Luka Govedič <[email protected]>
hmellor
added a commit
to hmellor/vllm
that referenced
this pull request
Oct 21, 2025
Signed-off-by: Harry Mellor <[email protected]>
hmellor
added a commit
that referenced
this pull request
Oct 22, 2025
Signed-off-by: Harry Mellor <[email protected]>
usberkeley
pushed a commit
to usberkeley/vllm
that referenced
this pull request
Oct 23, 2025
…m-project#27260) Signed-off-by: Harry Mellor <[email protected]>
albertoperdomo2
pushed a commit
to albertoperdomo2/vllm
that referenced
this pull request
Oct 23, 2025
…m-project#27260) Signed-off-by: Harry Mellor <[email protected]> Signed-off-by: Alberto Perdomo <[email protected]>
alhridoy
pushed a commit
to alhridoy/vllm
that referenced
this pull request
Oct 24, 2025
): name change compilation level to compilation mode, deprecation compilation level (vllm-project#26355) Signed-off-by: morrison-turnansky <[email protected]> Signed-off-by: Morrison Turnansky <[email protected]> Co-authored-by: Luka Govedič <[email protected]>
wangxiyuan
pushed a commit
to vllm-project/vllm-ascend
that referenced
this pull request
Oct 24, 2025
### What this PR does / why we need it? This is the step 1 of refactoring code to adapt with vllm main, and this pr aligned with vllm-project/vllm@17c540a 1. refactor deepseek to the latest code arch as of vllm-project/vllm@17c540a 2. bunches of fixes due to vllm changes - Fix `AscendScheduler` `__post_init__`, caused by vllm-project/vllm#25075 - Fix `AscendScheduler` init got an unexpected arg `block_size`, caused by vllm-project/vllm#26296 - Fix `KVCacheManager` `get_num_common_prefix_blocks` arg, caused by vllm-project/vllm#23485 - Fix `MLAAttention` import,caused by vllm-project/vllm#25103 - Fix `SharedFusedMoE` import, caused by vllm-project/vllm#26145 - Fix `LazyLoader` improt, caused by vllm-project/vllm#27022 - Fix `vllm.utils.swap_dict_values` improt, caused by vllm-project/vllm#26990 - Fix `Backend` enum import, caused by vllm-project/vllm#25893 - Fix `CompilationLevel` renaming to `CompilationMode` issue introduced by vllm-project/vllm#26355 - Fix fused_moe ops, caused by vllm-project/vllm#24097 - Fix bert model because of `inputs_embeds`, caused by vllm-project/vllm#25922 - Fix MRope because of `get_input_positions_tensor` to `get_mrope_input_positions`, caused by vllm-project/vllm#24172 - Fix `splitting_ops` changes introduced by vllm-project/vllm#25845 - Fix multi-modality changes introduced by vllm-project/vllm#16229 - Fix lora bias dropping issue introduced by vllm-project/vllm#25807 - Fix structured ouput break introduced by vllm-project/vllm#26737 ### Does this PR introduce _any_ user-facing change? ### How was this patch tested? CI passed with existing test. - vLLM version: v0.11.0rc3 - vLLM main: https://github.com/vllm-project/vllm/commit/v0.11.0 --------- Signed-off-by: MengqingCao <[email protected]> Signed-off-by: Icey <[email protected]> Co-authored-by: Icey <[email protected]>
xuebwang-amd
pushed a commit
to xuebwang-amd/vllm
that referenced
this pull request
Oct 24, 2025
): name change compilation level to compilation mode, deprecation compilation level (vllm-project#26355) Signed-off-by: morrison-turnansky <[email protected]> Signed-off-by: Morrison Turnansky <[email protected]> Co-authored-by: Luka Govedič <[email protected]> Signed-off-by: xuebwang-amd <[email protected]>
xuebwang-amd
pushed a commit
to xuebwang-amd/vllm
that referenced
this pull request
Oct 24, 2025
): name change compilation level to compilation mode, deprecation compilation level (vllm-project#26355) Signed-off-by: morrison-turnansky <[email protected]> Signed-off-by: Morrison Turnansky <[email protected]> Co-authored-by: Luka Govedič <[email protected]> Signed-off-by: xuebwang-amd <[email protected]>
kingsmad
pushed a commit
to kingsmad/vllm
that referenced
this pull request
Oct 25, 2025
…m-project#27260) Signed-off-by: Harry Mellor <[email protected]>
0xrushi
pushed a commit
to 0xrushi/vllm
that referenced
this pull request
Oct 26, 2025
): name change compilation level to compilation mode, deprecation compilation level (vllm-project#26355) Signed-off-by: morrison-turnansky <[email protected]> Signed-off-by: Morrison Turnansky <[email protected]> Co-authored-by: Luka Govedič <[email protected]> Signed-off-by: 0xrushi <[email protected]>
0xrushi
pushed a commit
to 0xrushi/vllm
that referenced
this pull request
Oct 26, 2025
…m-project#27260) Signed-off-by: Harry Mellor <[email protected]> Signed-off-by: 0xrushi <[email protected]>
0xrushi
pushed a commit
to 0xrushi/vllm
that referenced
this pull request
Oct 26, 2025
): name change compilation level to compilation mode, deprecation compilation level (vllm-project#26355) Signed-off-by: morrison-turnansky <[email protected]> Signed-off-by: Morrison Turnansky <[email protected]> Co-authored-by: Luka Govedič <[email protected]> Signed-off-by: 0xrushi <[email protected]>
0xrushi
pushed a commit
to 0xrushi/vllm
that referenced
this pull request
Oct 26, 2025
…m-project#27260) Signed-off-by: Harry Mellor <[email protected]> Signed-off-by: 0xrushi <[email protected]>
1 task
ilmarkov
pushed a commit
to neuralmagic/vllm
that referenced
this pull request
Nov 7, 2025
…m-project#27260) Signed-off-by: Harry Mellor <[email protected]>
rtourgeman
pushed a commit
to rtourgeman/vllm
that referenced
this pull request
Nov 10, 2025
): name change compilation level to compilation mode, deprecation compilation level (vllm-project#26355) Signed-off-by: morrison-turnansky <[email protected]> Signed-off-by: Morrison Turnansky <[email protected]> Co-authored-by: Luka Govedič <[email protected]>
rtourgeman
pushed a commit
to rtourgeman/vllm
that referenced
this pull request
Nov 10, 2025
…m-project#27260) Signed-off-by: Harry Mellor <[email protected]>
Zhathw
pushed a commit
to Zhathw/vllm
that referenced
this pull request
Nov 12, 2025
): name change compilation level to compilation mode, deprecation compilation level (vllm-project#26355) Signed-off-by: morrison-turnansky <[email protected]> Signed-off-by: Morrison Turnansky <[email protected]> Co-authored-by: Luka Govedič <[email protected]>
luolun
pushed a commit
to luolun/vllm-ascend
that referenced
this pull request
Nov 19, 2025
) ### What this PR does / why we need it? This is the step 1 of refactoring code to adapt with vllm main, and this pr aligned with vllm-project/vllm@17c540a 1. refactor deepseek to the latest code arch as of vllm-project/vllm@17c540a 2. bunches of fixes due to vllm changes - Fix `AscendScheduler` `__post_init__`, caused by vllm-project/vllm#25075 - Fix `AscendScheduler` init got an unexpected arg `block_size`, caused by vllm-project/vllm#26296 - Fix `KVCacheManager` `get_num_common_prefix_blocks` arg, caused by vllm-project/vllm#23485 - Fix `MLAAttention` import,caused by vllm-project/vllm#25103 - Fix `SharedFusedMoE` import, caused by vllm-project/vllm#26145 - Fix `LazyLoader` improt, caused by vllm-project/vllm#27022 - Fix `vllm.utils.swap_dict_values` improt, caused by vllm-project/vllm#26990 - Fix `Backend` enum import, caused by vllm-project/vllm#25893 - Fix `CompilationLevel` renaming to `CompilationMode` issue introduced by vllm-project/vllm#26355 - Fix fused_moe ops, caused by vllm-project/vllm#24097 - Fix bert model because of `inputs_embeds`, caused by vllm-project/vllm#25922 - Fix MRope because of `get_input_positions_tensor` to `get_mrope_input_positions`, caused by vllm-project/vllm#24172 - Fix `splitting_ops` changes introduced by vllm-project/vllm#25845 - Fix multi-modality changes introduced by vllm-project/vllm#16229 - Fix lora bias dropping issue introduced by vllm-project/vllm#25807 - Fix structured ouput break introduced by vllm-project/vllm#26737 ### Does this PR introduce _any_ user-facing change? ### How was this patch tested? CI passed with existing test. - vLLM version: v0.11.0rc3 - vLLM main: https://github.com/vllm-project/vllm/commit/v0.11.0 --------- Signed-off-by: MengqingCao <[email protected]> Signed-off-by: Icey <[email protected]> Co-authored-by: Icey <[email protected]> Signed-off-by: luolun <[email protected]>
hwhaokun
pushed a commit
to hwhaokun/vllm-ascend
that referenced
this pull request
Nov 19, 2025
) ### What this PR does / why we need it? This is the step 1 of refactoring code to adapt with vllm main, and this pr aligned with vllm-project/vllm@17c540a 1. refactor deepseek to the latest code arch as of vllm-project/vllm@17c540a 2. bunches of fixes due to vllm changes - Fix `AscendScheduler` `__post_init__`, caused by vllm-project/vllm#25075 - Fix `AscendScheduler` init got an unexpected arg `block_size`, caused by vllm-project/vllm#26296 - Fix `KVCacheManager` `get_num_common_prefix_blocks` arg, caused by vllm-project/vllm#23485 - Fix `MLAAttention` import,caused by vllm-project/vllm#25103 - Fix `SharedFusedMoE` import, caused by vllm-project/vllm#26145 - Fix `LazyLoader` improt, caused by vllm-project/vllm#27022 - Fix `vllm.utils.swap_dict_values` improt, caused by vllm-project/vllm#26990 - Fix `Backend` enum import, caused by vllm-project/vllm#25893 - Fix `CompilationLevel` renaming to `CompilationMode` issue introduced by vllm-project/vllm#26355 - Fix fused_moe ops, caused by vllm-project/vllm#24097 - Fix bert model because of `inputs_embeds`, caused by vllm-project/vllm#25922 - Fix MRope because of `get_input_positions_tensor` to `get_mrope_input_positions`, caused by vllm-project/vllm#24172 - Fix `splitting_ops` changes introduced by vllm-project/vllm#25845 - Fix multi-modality changes introduced by vllm-project/vllm#16229 - Fix lora bias dropping issue introduced by vllm-project/vllm#25807 - Fix structured ouput break introduced by vllm-project/vllm#26737 ### Does this PR introduce _any_ user-facing change? ### How was this patch tested? CI passed with existing test. - vLLM version: v0.11.0rc3 - vLLM main: https://github.com/vllm-project/vllm/commit/v0.11.0 --------- Signed-off-by: MengqingCao <[email protected]> Signed-off-by: Icey <[email protected]> Co-authored-by: Icey <[email protected]> Signed-off-by: hwhaokun <[email protected]>
NSDie
pushed a commit
to NSDie/vllm-ascend
that referenced
this pull request
Nov 24, 2025
) ### What this PR does / why we need it? This is the step 1 of refactoring code to adapt with vllm main, and this pr aligned with vllm-project/vllm@17c540a 1. refactor deepseek to the latest code arch as of vllm-project/vllm@17c540a 2. bunches of fixes due to vllm changes - Fix `AscendScheduler` `__post_init__`, caused by vllm-project/vllm#25075 - Fix `AscendScheduler` init got an unexpected arg `block_size`, caused by vllm-project/vllm#26296 - Fix `KVCacheManager` `get_num_common_prefix_blocks` arg, caused by vllm-project/vllm#23485 - Fix `MLAAttention` import,caused by vllm-project/vllm#25103 - Fix `SharedFusedMoE` import, caused by vllm-project/vllm#26145 - Fix `LazyLoader` improt, caused by vllm-project/vllm#27022 - Fix `vllm.utils.swap_dict_values` improt, caused by vllm-project/vllm#26990 - Fix `Backend` enum import, caused by vllm-project/vllm#25893 - Fix `CompilationLevel` renaming to `CompilationMode` issue introduced by vllm-project/vllm#26355 - Fix fused_moe ops, caused by vllm-project/vllm#24097 - Fix bert model because of `inputs_embeds`, caused by vllm-project/vllm#25922 - Fix MRope because of `get_input_positions_tensor` to `get_mrope_input_positions`, caused by vllm-project/vllm#24172 - Fix `splitting_ops` changes introduced by vllm-project/vllm#25845 - Fix multi-modality changes introduced by vllm-project/vllm#16229 - Fix lora bias dropping issue introduced by vllm-project/vllm#25807 - Fix structured ouput break introduced by vllm-project/vllm#26737 ### Does this PR introduce _any_ user-facing change? ### How was this patch tested? CI passed with existing test. - vLLM version: v0.11.0rc3 - vLLM main: https://github.com/vllm-project/vllm/commit/v0.11.0 --------- Signed-off-by: MengqingCao <[email protected]> Signed-off-by: Icey <[email protected]> Co-authored-by: Icey <[email protected]> Signed-off-by: nsdie <[email protected]>
devpatelio
pushed a commit
to SumanthRH/vllm
that referenced
this pull request
Nov 29, 2025
): name change compilation level to compilation mode, deprecation compilation level (vllm-project#26355) Signed-off-by: morrison-turnansky <[email protected]> Signed-off-by: Morrison Turnansky <[email protected]> Co-authored-by: Luka Govedič <[email protected]>
devpatelio
pushed a commit
to SumanthRH/vllm
that referenced
this pull request
Nov 29, 2025
…m-project#27260) Signed-off-by: Harry Mellor <[email protected]>
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Labels
documentation
Improvements or additions to documentation
frontend
llama
Related to Llama models
ready
ONLY add when PR is ready to merge/full CI is needed
speculative-decoding
tpu
Related to Google TPUs
v1
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Purpose
See #20283 (comment)
The purpose of this PR is to perform variable name changes and deprecation warnings for compilation level to compilation mode. Enum values names are also changed. No true changes of behavior nor breaking changes have occurred.
Test Plan
Test Result
Essential Elements of an Effective PR Description Checklist
supported_models.mdandexamplesfor a new model.