Skip to content

Commit 66e4d6d

Browse files
fix
Signed-off-by: baonudesifeizhai <[email protected]>
1 parent d2a30a2 commit 66e4d6d

File tree

2 files changed

+7
-1
lines changed

2 files changed

+7
-1
lines changed

vllm/attention/layers/cross_attention.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -37,6 +37,8 @@ def _get_cross_slot_mapping(encoder_seq_lens: np.ndarray,
3737
device: torch.device) -> torch.Tensor:
3838
"""Get cross-attention slot mappings."""
3939

40+
encoder_seq_lens = np.atleast_1d(encoder_seq_lens)
41+
4042
block_size = kv_cache_spec.block_size
4143
slot_mappings = []
4244

vllm/v1/worker/gpu_model_runner.py

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -886,7 +886,11 @@ def _get_encoder_seq_lens(
886886

887887
# Build encoder_seq_lens array mapping request indices to
888888
# encoder lengths for inputs scheduled in this batch
889-
encoder_seq_lens = np.zeros(num_reqs, dtype=np.int32)
889+
num_reqs_int = int(num_reqs)
890+
if num_reqs_int == 0:
891+
return np.zeros((0,), dtype=np.int32)
892+
893+
encoder_seq_lens = np.zeros((num_reqs_int,), dtype=np.int32)
890894
for req_id in scheduler_output.scheduled_encoder_inputs:
891895
req_index = self.input_batch.req_id_to_index[req_id]
892896
encoder_seq_lens[req_index] = self.max_encoder_len

0 commit comments

Comments
 (0)