mpt-7b-8k-chat-awq / flash_attn_triton.py

Commit History

MPT 7B 8K quantized
5c660fe

casperhansen commited on