Commit History
fix evals (#447)
ee26281
unverified
support user defined prompters, pretokenized datasets in config, local parquet, local arrow files (#348)
d2e7f27
unverified
disable eval using multipack for now (#437)
f733d0f
unverified
fix comma, not a tuple (#436)
008505c
unverified
use save_strategy from config if available (#434)
b3f5e00
unverified
set env for FSDP offload params (#433)
5247c50
unverified
Fix(config): Update handling of deepspeed config (#404)
c01015f
unverified
fix eval steps and strategy (#403)
da10af0
unverified
add utils.data.prepare_dataset
2e22404
use context manager to run things on rank0 before others (#397)
fc2d6be
unverified
don't use mask expansion for inference (#392)
1687be6
unverified
Feat(config): add max steps (#387)
3c2ad00
unverified
Added "epoch" evaluation_strategy (#388)
5d48a10
unverified
Feat(config): Add hub_strategy (#386)
73a0b6e
unverified
don't pass rope_scaling kwarg if it's None (#383)
919246f
unverified
Fix crash when running without CUDA
15f6e57
try to detect accelerate and only use device_map=None in that case (#373)
094fc2c
unverified
remove unnecessary local variable
0c96727
simplify `load_tokenizer`
efb3b2c
improve GPU logging to break out pytorch cache and system mem
7b55fe6
quiet noise from llama tokenizer by setting pad token earlier
e029ab3
extract module for working with cfg
8cec513
fix DefaultDict.__or__
a13e45d
Attention mask and position id fixes for packing (#285)
2bb0b78
unverified
Add wandb_entity to wandb options, update example configs, update README (#361)
7019509
unverified
Fix(model loading): Warn when model revision is passed to gptq (#364)
96bd6ae
unverified
Feat: Add rope scaling (#343)
b521206
unverified
Merge pull request #356 from tmm1/load_model-args
11ddccb
unverified
simplify load_model signature
7181022
log GPU memory usage
e303d64
ensure enable_input_require_grads is called on model before getting the peft model (#345)
176b888
unverified
experimental llama 2 chat support (#296)
3392270
unverified
Jan Philipp Harries
Jan Philipp Harries
commited on
optimize the iteration when tokenizeing large datasets (#332)
fe28543
unverified
fix typo
2eda9e0
scope flash-attn+qlora fix correctly, scope to llama, add comment
78b9efb
move flash-attn monkey patch alongside the others
312a9fa
ensure flash-attn fixes happen in both adapter/lora modes, and use torch_dtype
248bf90
qlora w flash attention fixes (#333)
77085ea
unverified
add peft install back since it doesn't get installed by setup.py (#331)
db2a358
unverified
don't resize embeddings to multiples of 32x by default
1066751
fix axolotl training args dataclass annotation
ebaec3c
Merge pull request #276 from theobjectivedad/logging_enhancement
6f16c45
unverified
Fixed pre-commit problems, fixed small bug in logging_config to handle LOG_LEVEL env var
b1f4f7a
Merge branch 'OpenAccess-AI-Collective:main' into logging_enhancement
83237b8
unverified
The Objective Dad
commited on