flash_attention + sample packing for stablelm 3b (#671) 2d60ba3 unverified winglian commited on Oct 5, 2023
fix for flash attn w mistral w/o sammple packing (#648) b2edaae unverified winglian commited on Sep 28, 2023
skip some flash attn patches unless explicitly enabled (#643) 895f0a0 unverified winglian commited on Sep 27, 2023
btlm and falcon monkey patches for flash attn (#566) 6b9b229 unverified winglian commited on Sep 17, 2023
Add training callback to send predictions to WandB table (#521) 5b67ea9 unverified Glavin001 commited on Sep 13, 2023
ReLoRA implementation (with quantization) (#322) bde3c5a unverified chargoddard winglian commited on Aug 24, 2023
fix eval regression caused in 13f7efaf74fcd3c4514277ccb71914c589873f6a a213d99 tmm1 commited on Aug 21, 2023
Attention mask and position id fixes for packing (#285) 2bb0b78 unverified winglian commited on Aug 12, 2023
Update XFormers Attention Monkeypatch to handle Llama-2 70B (GQA) (#339) 10405b9 unverified ssmi153 commited on Aug 6, 2023
fix sdp attention to use the flash/mem-efficient context manaager a032c9f winglian commited on Jul 20, 2023
Fixed pre-commit problems, fixed small bug in logging_config to handle LOG_LEVEL env var b1f4f7a theobjectivedad commited on Jul 15, 2023
Update src/axolotl/monkeypatch/llama_attn_hijack_xformers.py 1076bcb unverified winglian Nanobit commited on May 31, 2023
Update src/axolotl/monkeypatch/llama_attn_hijack_xformers.py 2daa683 unverified winglian Nanobit commited on May 31, 2023
copy xformers attn from ooba since we removed dep on alpaca_lora_4bit 6cb2310 winglian commited on May 31, 2023