Skip to content
Merged
Show file tree
Hide file tree
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 3 additions & 3 deletions benchmarks/benchmark_throughput.py
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@
import random
import time
from functools import cache
from typing import Dict, List, Optional, Tuple
from typing import Dict, List, Optional, Tuple, Union

import torch
import uvloop
Expand Down Expand Up @@ -171,7 +171,7 @@ def run_vllm(
llm = LLM(**dataclasses.asdict(engine_args))

# Add the requests to the engine.
prompts: List[TextPrompt | TokensPrompt] = []
prompts: List[Union[TextPrompt, TokensPrompt]] = []
sampling_params: List[SamplingParams] = []
for request in requests:
prompts.append(
Expand Down Expand Up @@ -232,7 +232,7 @@ async def run_vllm_async(
engine_args, disable_frontend_multiprocessing) as llm:

# Add the requests to the engine.
prompts: List[TextPrompt | TokensPrompt] = []
prompts: List[Union[TextPrompt, TokensPrompt]] = []
sampling_params: List[SamplingParams] = []
lora_requests: List[Optional[LoRARequest]] = []
for request in requests:
Expand Down
4 changes: 3 additions & 1 deletion vllm/engine/arg_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -255,7 +255,9 @@ def add_cli_args(parser: FlexibleArgumentParser) -> FlexibleArgumentParser:
parser.add_argument(
'--skip-tokenizer-init',
action='store_true',
help='Skip initialization of tokenizer and detokenizer.')
help='Skip tokenizer\detokenizer use during model inference. '
'Use of tokenizer\detokenizer increases host overhead and might '
'end up of GPU underutilization')
parser.add_argument(
'--revision',
type=nullable_str,
Expand Down