Skip to content
Open
Show file tree
Hide file tree
Changes from 3 commits
Commits
Show all changes
23 commits
Select commit Hold shift + click to select a range
66e4d6d
fix
baonudesifeizhai Sep 18, 2025
1b215ee
yapf
baonudesifeizhai Sep 18, 2025
1f8c6e3
Merge branch 'main' into whisper-cudagraphs-support
baonudesifeizhai Sep 18, 2025
4fda4fc
Update vllm/v1/worker/gpu_model_runner.py
baonudesifeizhai Sep 18, 2025
e8d12d7
yapf
baonudesifeizhai Sep 18, 2025
88059d3
fix
baonudesifeizhai Sep 19, 2025
28800dc
fix and debug
baonudesifeizhai Sep 19, 2025
51e8742
yapf fix
baonudesifeizhai Sep 19, 2025
573ffcd
Merge branch 'main' into whisper-cudagraphs-support
baonudesifeizhai Sep 19, 2025
ebf1d39
Merge branch 'main' into whisper-cudagraphs-support
baonudesifeizhai Sep 19, 2025
1195357
add log
baonudesifeizhai Sep 19, 2025
95954c1
Merge branch 'whisper-cudagraphs-support' of https://github.com/baonu…
baonudesifeizhai Sep 19, 2025
b28a68a
remove logger
baonudesifeizhai Sep 19, 2025
59583c0
Merge branch 'vllm-project:main' into whisper-cudagraphs-support
baonudesifeizhai Sep 19, 2025
83bdfc3
Merge branch 'vllm-project:main' into whisper-cudagraphs-support
baonudesifeizhai Sep 22, 2025
47f9cd4
Merge branch 'main' into whisper-cudagraphs-support
baonudesifeizhai Sep 23, 2025
342f5a8
Merge branch 'main' into whisper-cudagraphs-support
baonudesifeizhai Sep 24, 2025
89af950
Merge branch 'vllm-project:main' into whisper-cudagraphs-support
baonudesifeizhai Sep 25, 2025
6092e13
Merge branch 'vllm-project:main' into whisper-cudagraphs-support
baonudesifeizhai Sep 26, 2025
d8eb97d
Merge branch 'vllm-project:main' into whisper-cudagraphs-support
baonudesifeizhai Oct 11, 2025
2f4e230
fix format error and add tracker in encoder lengths
baonudesifeizhai Oct 11, 2025
f8dd813
Merge branch 'main' into whisper-cudagraphs-support
baonudesifeizhai Oct 16, 2025
159c66f
Merge branch 'main' into whisper-cudagraphs-support
baonudesifeizhai Oct 16, 2025
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 2 additions & 0 deletions vllm/attention/layers/cross_attention.py
Original file line number Diff line number Diff line change
Expand Up @@ -37,6 +37,8 @@ def _get_cross_slot_mapping(encoder_seq_lens: np.ndarray,
device: torch.device) -> torch.Tensor:
"""Get cross-attention slot mappings."""

encoder_seq_lens = np.atleast_1d(encoder_seq_lens)

block_size = kv_cache_spec.block_size
slot_mappings = []

Expand Down
6 changes: 5 additions & 1 deletion vllm/v1/worker/gpu_model_runner.py
Original file line number Diff line number Diff line change
Expand Up @@ -886,7 +886,11 @@ def _get_encoder_seq_lens(

# Build encoder_seq_lens array mapping request indices to
# encoder lengths for inputs scheduled in this batch
encoder_seq_lens = np.zeros(num_reqs, dtype=np.int32)
num_reqs_int = int(num_reqs)
if num_reqs_int == 0:
return np.zeros((0, ), dtype=np.int32)

encoder_seq_lens = np.zeros((num_reqs_int, ), dtype=np.int32)
for req_id in scheduler_output.scheduled_encoder_inputs:
req_index = self.input_batch.req_id_to_index[req_id]
encoder_seq_lens[req_index] = self.max_encoder_len
Expand Down
Loading