Enable flash_attention_2 support since the underlying Mistral model supports it 93dbd18 verified winglian commited on Apr 20