We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 3ccf00b commit afd0cc4Copy full SHA for afd0cc4
src/transformers/models/blt/modeling_blt.py
@@ -750,7 +750,7 @@ def forward(
750
751
if i == 0 or self.config.cross_attn_all_layers:
752
# Use cross attention to extract info from patch_embeds into hidden_states
753
- cross_attention_output, _, _ = self.cross_attn_layers[i](
+ cross_attention_output, _ = self.cross_attn_layers[i](
754
hidden_states=hidden_states,
755
cross_attention_states=patch_embeds,
756
attention_mask=cross_mask,
0 commit comments