qwerrwe / tests

Commit History

feat: validate sample packing requires flash_attention (#1465)
bf4cd67
unverified

Nanobit commited on

Support loading datasets saved via save_to_disk (#1432)
e634118
unverified

fozziethebeat commited on

make sure to capture non-null defaults from config validation (#1415)
601b77b
unverified

winglian commited on

fix(dataset): normalize tokenizer config and change hash from tokenizer class to tokenizer path (#1298)
ff939d8
unverified

Nanobit commited on

strip out hacky qlora-fsdp workarounds now that qlora-fsdp fixes are upstreamed (#1428)
2a1589f
unverified

winglian commited on

Feat: Add sharegpt multirole (#1137)
40a88e8
unverified

Nanobit commited on

ORPO (#1419)
2ea70eb
unverified

winglian commited on

Train parameters exclusively in specific ranges (#1390)
05bcc9e
unverified

seungduk commited on

Add Glaive conversation format support (#1365)
b7d8a7d
unverified

Brian Fitzgerald winglian commited on

plain input/output prompt strategy w/o chat templates (#1346)
4d09b42
unverified

winglian commited on

run tests again on Modal (#1289) [skip ci]
0001862
unverified

winglian commited on

fix for protected model_ namespace w pydantic (#1345)
6b3b271
unverified

winglian commited on

more fixes 20240228 (#1342) [skip ci]
0f985e1
unverified

winglian commited on

Pydantic 2.x cfg (#1239)
cc3cebf
unverified

winglian commited on

make mlflow optional (#1317)
5894f0e
unverified

winglian commited on

Scheduler implementation of Continual Pre-Training of Large Language Models: How to (re)warm your model? (#1273)
8430db2
unverified

jinwonkim93 commited on

Pretrain transforms (#1261)
c7cf381
unverified

winglian commited on

relora: magnitude pruning of the optimizer (#1245)
8c2e05a
unverified

winglian commited on

support for true batches with multipack (#1230)
00568c1
unverified

winglian commited on

Support for additional_special_tokens (#1221) [skip ci]
25e037f
unverified

DreamGenX winglian commited on

Peft lotfq (#1222)
4cb7900
unverified

winglian commited on

ADD: warning if hub_model_id ist set but not any save strategy (#1202)
af29d81
unverified

JohanWork winglian commited on

Feat/chatml add system message (#1117)
98b4762
unverified

mhenrichsen Mads Henrichsen winglian commited on

Phi2 multipack (#1173)
814aee6
unverified

winglian commited on

Feat(test): Add tests for alpaca chatml prompt tokenizer (#1088)
5439707
unverified

JohanWork Nanobit commited on

Falcon embeddings (#1149) [skip docker]
e799e08
unverified

winglian commited on

set fp16 to false if bf16, update bf16: auto in example YAMLs (#1122) [skip ci]
782b6a4
unverified

winglian Nanobit commited on

Deprecate max packed sequence len (#1141)
2ce5c0d
unverified

winglian commited on

Multipack simplify for Mixtral (#1142)
6910e6a
unverified

winglian commited on

Add shifted sparse attention (#973) [skip-ci]
1d70f24
unverified

jrc joecummings winglian commited on

Add `layers_to_transform` for `lora_config` (#1118)
8487b97
unverified

xzuyn commited on

Enable or disable bf16 support based on availability (#1116)
0865613
unverified

Simon Hällqvist commited on

keep gate in fp32 for 16 bit loras (#1105)
da97285
unverified

winglian commited on

add gptneox embeddings, fix phi2 inputs, also fix the casting (#1083)
78c5b19
unverified

winglian commited on

update sharegpt conversations when chatml chat template is set (#1075) [skip ci]
0ce1a65
unverified

winglian commited on

fix: `train_on_inputs: true` ignored for sharegpt (#1045) [skip ci]
043c386
unverified

Nanobit winglian commited on

be more robust about checking embedding modules for lora finetunes (#1074) [skip ci]
0f10080
unverified

winglian commited on

attempt to also run e2e tests that needs gpus (#1070)
788649f
unverified

winglian commited on

fix double eos token for chatml (#1054) [skip ci]
651b7a3
unverified

winglian commited on

Phi2 rewrite (#1058)
732851f
unverified

winglian commited on

streaming multipack for pretraining dataset (#959)
553c80f
unverified

jinwonkim93 jinwonkim93@github.com winglian commited on

RL/DPO (#935)
f243c21

winglian commited on

bump transformers and update attention class map name (#1023)
bcc78d8
unverified

winglian commited on

Feat: Warns to add to modules_to_save when adding tokens or switching special_tokens (#787)
1ffa386
unverified

Nanobit commited on

fix mistral prompt assembly (#982)
7bbaac9
unverified

hamel commited on

Fix prompt assembly for llama (#952)
5ada140
unverified

hamel tokestermw commited on

Respect sequence_len in config for `type: llama2_chat` (#926)
f1de29d
unverified

hamel commited on

support for mamba (#915)
40a6362
unverified

winglian commited on

Feat(wandb): Refactor to be more flexible (#767)
a1da39c
unverified

Nanobit commited on