Skip to content

Commit c61ec83

Browse files
DarkLight1337DefTruth
authored andcommitted
Revert "[Bugfix] Limit profiling run sequence length by max_model_len (vllm-project#14785) (vllm-project#14892)
Signed-off-by: DarkLight1337 <[email protected]> Signed-off-by: DefTruth <[email protected]>
1 parent ca7d46b commit c61ec83

File tree

5 files changed

+0
-9
lines changed

5 files changed

+0
-9
lines changed

vllm/inputs/registry.py

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -330,11 +330,6 @@ def dummy_data_for_profiling(
330330
from vllm.multimodal import MultiModalKwargs
331331
from vllm.multimodal.profiling import MultiModalProfiler
332332

333-
if seq_len > model_config.max_model_len:
334-
raise AssertionError(
335-
f"Profiling attempted with sequence length ({seq_len}) "
336-
f"greater than model length ({model_config.max_model_len})")
337-
338333
if mm_registry.has_processor(model_config):
339334
tokenizer = cached_tokenizer_from_config(model_config)
340335
processor = mm_registry.create_processor(model_config,

vllm/worker/enc_dec_model_runner.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -281,7 +281,6 @@ def profile_run(self) -> None:
281281
for group_id in range(max_num_seqs):
282282
seq_len = (max_num_batched_tokens // max_num_seqs +
283283
(group_id < max_num_batched_tokens % max_num_seqs))
284-
seq_len = min(seq_len, self.model_config.max_model_len)
285284
batch_size += seq_len
286285

287286
decoder_dummy_data = self.input_registry \

vllm/worker/model_runner.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1302,7 +1302,6 @@ def _dummy_run(self,
13021302
for group_id in range(max_num_seqs):
13031303
seq_len = (max_num_batched_tokens // max_num_seqs +
13041304
(group_id < max_num_batched_tokens % max_num_seqs))
1305-
seq_len = min(seq_len, self.model_config.max_model_len)
13061305
batch_size += seq_len
13071306

13081307
dummy_data = self.input_registry \

vllm/worker/openvino_model_runner.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -148,7 +148,6 @@ def _prepare_model_input(
148148
seq_len = min(
149149
seq_data.get_len(),
150150
computed_len + seq_group_metadata.token_chunk_size,
151-
self.model_config.max_model_len,
152151
)
153152
if is_prompt:
154153
tokens = seq_data.get_token_ids()[computed_len:seq_len]

vllm/worker/xpu_model_runner.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -466,7 +466,6 @@ def profile_run(self) -> None:
466466
for group_id in range(max_num_seqs):
467467
seq_len = (max_num_batched_tokens // max_num_seqs +
468468
(group_id < max_num_batched_tokens % max_num_seqs))
469-
seq_len = min(seq_len, self.model_config.max_model_len)
470469
batch_size += seq_len
471470

472471
dummy_data = self.input_registry \

0 commit comments

Comments
 (0)