Commit History

ignore: address pr review
d03887f
unverified

Maxime commited on

fix: inference did not move the model to the correct device (#483)
17605b8
unverified

Maxime commited on

ignore: linter
a184549
unverified

Maxime commited on

fix: finetune model inference needs the dtype fix to work with flash-attn
f311df9
unverified

Maxime commited on

Fix missing 'packaging' wheel (#482)
c500d02
unverified

Maxime commited on

fix checkpints on multigpu (#481)
31f3e71
unverified

winglian commited on

Merge pull request #484 from OpenAccess-AI-Collective/reqs
56c4a94
unverified

tmm1 commited on

allow newer deps
c29117a

tmm1 commited on

fix types w lora (#478)
0b7ba57
unverified

winglian commited on

Fix(tokenizer): Fix condition to add pad token (#477)
71bd062
unverified

Nanobit commited on

improve llama pad token handling (#475)
cb9797e
unverified

winglian commited on

ReLoRA implementation (with quantization) (#322)
bde3c5a
unverified

chargoddard winglian commited on

Fix(doc): Clarify config (#466)
55c23c7
unverified

Nanobit commited on

workaround so training doesn't hang when packed dataloader batches aren't even (#461)
c69faee
unverified

winglian commited on

fix test fixture b/c hf trainer tokenization changed (#464)
d5dcf9c
unverified

winglian commited on

feat: add Metharme prompt strategy (#446)
f474650
unverified

TearGosling Nanobit commited on

recast loralayer, norm, lmhead + embed token weights per original qlora (#393)
96deb6b
unverified

winglian commited on

always drop samples that are too long (#452)
50682a3
unverified

winglian commited on

set env var for FSDP layer to wrap (#453)
5a1985b
unverified

winglian commited on

Merge pull request #451 from OpenAccess-AI-Collective/eval-is-causal
5e9c6af
unverified

tmm1 commited on

is_causal fix for evals?
fbf49a4

winglian commited on

add missing positional arg (#450)
58cf7e7
unverified

winglian commited on

feat(docs): improve user customized prompts (#443)
04a42b6
unverified

Nanobit commited on

feat(doc): add pillow to lambda instructions (#445)
919f4ca
unverified

Nanobit commited on

fix evals (#447)
ee26281
unverified

winglian commited on

gracefully handle empty input (#442)
9d629d8
unverified

winglian commited on

support user defined prompters, pretokenized datasets in config, local parquet, local arrow files (#348)
d2e7f27
unverified

winglian commited on

docs(readme): add `cd axolotl` (#440)
d21318d
unverified

philpax commited on

disable eval using multipack for now (#437)
f733d0f
unverified

winglian commited on

fix comma, not a tuple (#436)
008505c
unverified

winglian commited on

use save_strategy from config if available (#434)
b3f5e00
unverified

winglian commited on

set env for FSDP offload params (#433)
5247c50
unverified

winglian commited on

flash attn pip install (#426)
cf66547
unverified

mhenrichsen Ubuntu mhenrichsen Mads Henrichsen winglian commited on

standardize attn hijack patches (#381)
06edf17
unverified

tmm1 winglian commited on

remove extra accelearate in requirements (#430)
82e111a
unverified

winglian commited on

fix fixture for new tokenizer handling in transformers (#428)
8cace80
unverified

winglian commited on

fix orca prompts (#422)
1b7e860
unverified

winglian commited on

Fix(docs): Remove gptq+lora and fix xformer compat list (#423)
3d1f203
unverified

Nanobit commited on

just resort to tags ans use main-latest (#424)
d3d6fd6
unverified

winglian commited on

Fix(template): Inform to place stack trace to Issue (#417)
b7449a9
unverified

Nanobit winglian commited on

use inputs for image rather than outputs for docker metadata (#420)
5f80b35
unverified

winglian commited on

hopefully improve the README (#419)
2495909
unverified

winglian commited on

tag with latest as well for axolotl-runpod (#418)
7af8166
unverified

winglian commited on

Merge pull request #413 from mhenrichsen/chore/update-deepseed-config
f806e86
unverified

mhenrichsen commited on

Feat(doc): Add lr_quadratic_warmup to readme (#412)
2b990eb
unverified

Nanobit commited on

update path to align with fsdp example
bd8cab4

mhenrichsen commited on

Fix(config): Update handling of deepspeed config (#404)
c01015f
unverified

Nanobit commited on

Fix(docs): Update flash attn requirements (#409)
72fe3f8
unverified

Nanobit commited on