[Doc] Fix small typo in Transformers fallback (#14791)
Signed-off-by: Chen Zhang <zhangch99@outlook.com>
This commit is contained in:
parent
3fb17d26c8
commit
60c872d4b6
@ -101,7 +101,7 @@ class MyAttention(nn.Module):
|
|||||||
|
|
||||||
def forward(self, hidden_states, **kwargs): # <- kwargs are required
|
def forward(self, hidden_states, **kwargs): # <- kwargs are required
|
||||||
...
|
...
|
||||||
attention_interface = attention_interface = ALL_ATTENTION_FUNCTIONS[self.config._attn_implementation]
|
attention_interface = ALL_ATTENTION_FUNCTIONS[self.config._attn_implementation]
|
||||||
attn_output, attn_weights = attention_interface(
|
attn_output, attn_weights = attention_interface(
|
||||||
self,
|
self,
|
||||||
query_states,
|
query_states,
|
||||||
|
Loading…
x
Reference in New Issue
Block a user