Skip to content

Commit beb6246

Browse files
authored
[rollout,vllm] fix: max_num_seqs not take effect (#2960)
1 parent 980b018 commit beb6246

File tree

3 files changed

+3
-0
lines changed

3 files changed

+3
-0
lines changed

tests/workers/rollout/rollout_vllm/test_vllm_model_rope_scaling.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -44,6 +44,7 @@ def test_vllm_rollout_with_yarn_position_embeddings():
4444
"free_cache_engine": False,
4545
"disable_log_stats": True,
4646
"max_model_len": 35000 + 512,
47+
"max_num_seqs": 1024,
4748
"load_format": "auto",
4849
"val_kwargs": {
4950
"top_k": -1,

verl/workers/rollout/vllm_rollout/vllm_async_server.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -266,6 +266,7 @@ async def init_engine(self):
266266
disable_custom_all_reduce=True,
267267
skip_tokenizer_init=False,
268268
max_model_len=self.max_model_len,
269+
max_num_seqs=config.max_num_seqs,
269270
load_format="auto",
270271
disable_log_stats=config.disable_log_stats,
271272
max_num_batched_tokens=max_num_batched_tokens,

verl/workers/rollout/vllm_rollout/vllm_rollout_spmd.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -188,6 +188,7 @@ def __init__(self, model_path: str, config: DictConfig, tokenizer, model_hf_conf
188188
disable_custom_all_reduce=True,
189189
skip_tokenizer_init=False,
190190
max_model_len=max_model_len,
191+
max_num_seqs=config.max_num_seqs,
191192
load_format=load_format,
192193
disable_log_stats=config.disable_log_stats,
193194
max_num_batched_tokens=max_num_batched_tokens,

0 commit comments

Comments
 (0)