Nanobit commited on
Commit
e8717d3
1 Parent(s): 54c3b5b

Remove disable

Browse files
Files changed (1) hide show
  1. src/axolotl/flash_attn.py +1 -1
src/axolotl/flash_attn.py CHANGED
@@ -15,7 +15,7 @@ from flash_attn.flash_attn_interface import flash_attn_unpadded_qkvpacked_func
15
  from flash_attn.bert_padding import unpad_input, pad_input
16
 
17
 
18
- def forward( # pylint: disable=too-many-arguments
19
  self,
20
  hidden_states: torch.Tensor,
21
  attention_mask: Optional[torch.Tensor] = None,
 
15
  from flash_attn.bert_padding import unpad_input, pad_input
16
 
17
 
18
+ def forward(
19
  self,
20
  hidden_states: torch.Tensor,
21
  attention_mask: Optional[torch.Tensor] = None,