From accfee56d8988cae60915486310362db5831b1bd Mon Sep 17 00:00:00 2001 From: Gustavo de Rosa Date: Tue, 16 Jan 2024 16:06:28 +0000 Subject: [PATCH] Update modeling_phi.py --- modeling_phi.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/modeling_phi.py b/modeling_phi.py index 9f59e97..4b51f34 100644 --- a/modeling_phi.py +++ b/modeling_phi.py @@ -506,7 +506,7 @@ class PhiFlashAttention2(PhiAttention): value_states = value_states.to(target_dtype) attn_output = self._flash_attention_forward( - query_states, key_states, value_states, attention_mask, q_len, dropout=attn_dropout, softmax_scale=1.0 + query_states, key_states, value_states, attention_mask, q_len, dropout=attn_dropout, softmax_scale=None ) attn_output = attn_output.reshape(bsz, q_len, self.hidden_size).contiguous()