We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent ba33e88 commit 1004205Copy full SHA for 1004205
vllm/model_executor/models/deepseek_mtp.py
@@ -97,7 +97,7 @@ def forward(
97
) -> torch.Tensor:
98
assert inputs_embeds is not None
99
# masking inputs at position 0, as not needed by MTP
100
- inputs_embeds[positions == 0] = 0
+ inputs_embeds = torch.where(positions.unsqueeze(-1) == 0, 0, inputs_embeds)
101
inputs_embeds = self.enorm(inputs_embeds)
102
previous_hidden_states = self.hnorm(previous_hidden_states)
103
0 commit comments