Hugging Face
Models
Datasets
Spaces
Posts
Docs
Solutions
Pricing
Log In
Sign Up
Hellisotherpeople
/
toxic-dpo-v2-mistral7b-lora
like
2
PEFT
Safetensors
unalignment/toxic-dpo-v0.2
English
Not-For-All-Audiences
License:
mit
Model card
Files
Files and versions
Community
Use this model
8660312
toxic-dpo-v2-mistral7b-lora
Commit History
Upload adapter_config.json
8660312
verified
Hellisotherpeople
commited on
Feb 21
Upload special_tokens_map.json
f46bf94
verified
Hellisotherpeople
commited on
Feb 21
Upload tokenizer.model
3543817
verified
Hellisotherpeople
commited on
Feb 21
Upload tokenizer.json
922cec4
verified
Hellisotherpeople
commited on
Feb 21
Upload training_args.bin
1f63213
verified
Hellisotherpeople
commited on
Feb 21
Upload optimizer.pt
c6e1667
verified
Hellisotherpeople
commited on
Feb 21
Upload scheduler.pt
e996bbf
verified
Hellisotherpeople
commited on
Feb 21
Upload rng_state.pth
6d75ab7
verified
Hellisotherpeople
commited on
Feb 21
Upload trainer_state.json
2dff770
verified
Hellisotherpeople
commited on
Feb 21
initial commit
478e5a5
verified
Hellisotherpeople
commited on
Feb 21