qwerrwe / scripts

Commit History

improve GPU logging to break out pytorch cache and system mem
7b55fe6

tmm1 commited on

extract module for working with cfg
8cec513

tmm1 commited on

Attention mask and position id fixes for packing (#285)
2bb0b78
unverified

winglian commited on

Fix(save): Save as safetensors (#363)
a276c9c
unverified

Nanobit commited on

feat(merge): save tokenizer on merge (#362)
289d5c4
unverified

Nanobit commited on

Merge pull request #356 from tmm1/load_model-args
11ddccb
unverified

tmm1 commited on

simplify load_model signature
7181022

tmm1 commited on

log GPU memory usage
e303d64

tmm1 commited on

fix FSDP save of final model (#329)
894cba0
unverified

winglian commited on

add runpod envs to .bashrc, fix bnb env (#316)
cf62cfd
unverified

winglian commited on

misc fixes
d75adb9

winglian commited on

Fixed pre-commit problems, fixed small bug in logging_config to handle LOG_LEVEL env var
b1f4f7a

theobjectivedad commited on

Merge pull request #92 from OpenAccess-AI-Collective/flash-optimum
16bb627
unverified

winglian commited on

chore: Refactor inf_kwargs out
dc77c8e

Nanobit commited on

Merge branch 'main' into flash-optimum
fd2c981
unverified

winglian commited on

Merge pull request #177 from NanoCode012/fix/landmark-patch
8002ffb
unverified

winglian commited on

Merge pull request #159 from AngainorDev/patch-1
8e568bb
unverified

Nanobit commited on

Fix strict and Lint
b565ecf

Angainor commited on

Fix set mem_id for inference and refactor
974dc00

Nanobit commited on

Set mem cache args on inference
572d114

Nanobit commited on

fix formatting
958da70

winglian commited on

pass a prompt in from stdin for inference
c4e4f81

winglian commited on

Update scripts/finetune.py
759e867
unverified

winglian Nanobit commited on

address PR feedback
0c6f928

winglian commited on

add streaming dataset support for pretraining datasets
eea2731

winglian commited on

more tweaks to do pre-training with bettertransformers
1210dc8

winglian commited on

experimental expansion of ctx len
488a67d

winglian commited on

add flash attn context for efficient training and attempt setting model to train mode:
8792199

winglian commited on

add support for opimum bettertransformers
1edc30c

winglian commited on

Merge branch 'main' into patch-1
79e2a6f
unverified

Angainor Development commited on

Remove explicit definition of cfg.inference
c250898
unverified

Angainor Development commited on

formatting for linter
f36e227
unverified

winglian commited on

Add streaming inference & fix stopping at EOS
fec6bcc

Glavin001 commited on

Feed cfg.inference
bd3b537
unverified

Angainor Development commited on

Set matmul tf32
52765ac

Nanobit commited on

new prompters, misc fixes for output dir missing using fsdp, and changing max seq len
4ac9e25

winglian commited on

fix device map
74ebbf4

winglian commited on

fix batch size calculation
5a631b3

winglian commited on

Merge pull request #119 from NanoCode012/feat/update-inference
fac4600
unverified

Nanobit commited on

Increase max_new_tokens
33d4017
unverified

Nanobit winglian commited on

Merge pull request #120 from OpenAccess-AI-Collective/model-from-path
c7021e1
unverified

winglian commited on

black formatting
6fa40bf

winglian commited on

add support for gradient accumulation steps
3aad5f3

winglian commited on

fix up tokenizer config, isort fix
39a208c

winglian commited on

Feat: Swap to GenerationConfig
988aeb9

Nanobit commited on

Merge pull request #108 from OpenAccess-AI-Collective/docker-gptq
bbc5bc5
unverified

winglian commited on

Fix security issue or ignore false positives
a1f9850

Nanobit commited on

Apply isort then black
37293dc

Nanobit commited on

Delete extract_lora.py
96e8378

Nanobit commited on