Skip to content

Commit d894ee0

Browse files
committed
fix the doc
1 parent 3bf3ae4 commit d894ee0

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

src/transformers/modeling_flash_attention_utils.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -564,7 +564,7 @@ def _flash_attention_forward(
564564
attention_mask (`torch.Tensor`, *optional*):
565565
The padding mask - corresponds to a tensor of size `(batch_size, seq_len)` where 0 stands for the
566566
position of padding tokens and 1 for the position of non-padding tokens.
567-
implementation (`str`, *optional*):
567+
attb_implementation (`str`, *optional*):
568568
The attention implementation to use. If None, will default to the one based on the environment.
569569
"""
570570
(flash_fn, flash_varlen_fn, pad_fn, unpad_fn), process_flash_kwargs_fn = lazy_import_flash_attention(

0 commit comments

Comments
 (0)