mirror of
https://github.com/saymrwulf/transformers.git
synced 2026-05-14 20:58:08 +00:00
remove layer_idx
This commit is contained in:
parent
caaa5e5508
commit
5060a334de
1 changed files with 0 additions and 1 deletions
|
|
@ -269,7 +269,6 @@ class LlamaAttention(nn.Module):
|
|||
if self.config._attn_implementation != "eager":
|
||||
attention_interface = ALL_ATTENTION_FUNCTIONS[self.config._attn_implementation]
|
||||
|
||||
kwargs["layer_idx"] = self.layer_idx
|
||||
attn_output, attn_weights = attention_interface(
|
||||
self,
|
||||
query_states,
|
||||
|
|
|
|||
Loading…
Reference in a new issue