Skip to content

Commit 932df1a

Browse files
SmartManojshreyankg
authored andcommitted
[Misc] Set default value of seed to None (vllm-project#14274)
Signed-off-by: மனோஜ்குமார் பழனிச்சாமி <[email protected]>
1 parent 0294bce commit 932df1a

File tree

9 files changed

+15
-7
lines changed

9 files changed

+15
-7
lines changed

tests/distributed/test_torchrun_example.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,8 @@
2525
tensor_parallel_size=2,
2626
distributed_executor_backend="external_launcher",
2727
gpu_memory_utilization=random.uniform(0.7, 0.9),
28-
swap_space=random.randint(1, 4))
28+
swap_space=random.randint(1, 4),
29+
seed=0)
2930

3031
outputs = llm.generate(prompts, sampling_params)
3132

tests/entrypoints/llm/test_encode.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,8 @@ def llm():
3434
max_num_batched_tokens=32768,
3535
tensor_parallel_size=1,
3636
gpu_memory_utilization=0.75,
37-
enforce_eager=True)
37+
enforce_eager=True,
38+
seed=0)
3839

3940
with llm.deprecate_legacy_api():
4041
yield weakref.proxy(llm)

tests/entrypoints/llm/test_guided_generate.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@
2121
def llm():
2222
# pytest caches the fixture so we use weakref.proxy to
2323
# enable garbage collection
24-
llm = LLM(model=MODEL_NAME, max_model_len=1024)
24+
llm = LLM(model=MODEL_NAME, max_model_len=1024, seed=0)
2525

2626
with llm.deprecate_legacy_api():
2727
yield weakref.proxy(llm)

tests/entrypoints/openai/test_chat_echo.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -24,6 +24,8 @@ def server():
2424
"4080",
2525
"--chat-template",
2626
DUMMY_CHAT_TEMPLATE,
27+
"--seed",
28+
"0",
2729
]
2830

2931
with RemoteOpenAIServer(MODEL_NAME, args) as remote_server:

tests/entrypoints/openai/test_metrics.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -47,6 +47,8 @@ def default_server_args():
4747
"--enforce-eager",
4848
"--max-num-seqs",
4949
"128",
50+
"--seed",
51+
"0",
5052
]
5153

5254

tests/entrypoints/openai/test_root_path.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -30,6 +30,8 @@ def server():
3030
"/" + ROOT_PATH,
3131
"--chat-template",
3232
DUMMY_CHAT_TEMPLATE,
33+
"--seed",
34+
"0",
3335
]
3436
envs = os.environ.copy()
3537

vllm/engine/arg_utils.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -105,7 +105,7 @@ class EngineArgs:
105105
config_format: ConfigFormat = ConfigFormat.AUTO
106106
dtype: str = 'auto'
107107
kv_cache_dtype: str = 'auto'
108-
seed: int = 0
108+
seed: Optional[int] = None
109109
max_model_len: Optional[int] = None
110110
# Note: Specifying a custom executor backend by passing a class
111111
# is intended for expert use only. The API may change without

vllm/entrypoints/llm.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -169,7 +169,7 @@ def __init__(
169169
quantization: Optional[str] = None,
170170
revision: Optional[str] = None,
171171
tokenizer_revision: Optional[str] = None,
172-
seed: int = 0,
172+
seed: Optional[int] = None,
173173
gpu_memory_utilization: float = 0.9,
174174
swap_space: float = 4,
175175
cpu_offload_gb: float = 0,

vllm/utils.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -644,7 +644,7 @@ def create_kv_caches_with_random_flash(
644644
head_size: int,
645645
cache_dtype: Optional[Union[str, torch.dtype]],
646646
model_dtype: Optional[Union[str, torch.dtype]] = None,
647-
seed: int = 0,
647+
seed: Optional[int] = None,
648648
device: Optional[str] = "cuda",
649649
) -> tuple[list[torch.Tensor], list[torch.Tensor]]:
650650
from vllm.platforms import current_platform
@@ -681,7 +681,7 @@ def create_kv_caches_with_random(
681681
head_size: int,
682682
cache_dtype: Optional[Union[str, torch.dtype]],
683683
model_dtype: Optional[Union[str, torch.dtype]] = None,
684-
seed: int = 0,
684+
seed: Optional[int] = None,
685685
device: Optional[str] = "cuda",
686686
) -> tuple[list[torch.Tensor], list[torch.Tensor]]:
687687

0 commit comments

Comments
 (0)