datasets: | |
- allenai/ai2_arc | |
- unalignment/spicy-3.1 | |
- codeparrot/apps | |
- facebook/belebele | |
- boolq | |
- jondurbin/cinematika-v0.1 | |
- drop | |
- lmsys/lmsys-chat-1m | |
- TIGER-Lab/MathInstruct | |
- cais/mmlu | |
- Muennighoff/natural-instructions | |
- openbookqa | |
- piqa | |
- Vezora/Tested-22k-Python-Alpaca | |
- cakiki/rosetta-code | |
- Open-Orca/SlimOrca | |
- spider | |
- squad_v2 | |
- migtissera/Synthia-v1.3 | |
- datasets/winogrande | |
- nvidia/HelpSteer | |
- Intel/orca_dpo_pairs | |
- unalignment/toxic-dpo-v0.1 | |
- jondurbin/truthy-dpo-v0.1 | |
- allenai/ultrafeedback_binarized_cleaned | |
- Squish42/bluemoon-fandom-1-1-rp-cleaned | |
- LDJnr/Capybara | |
- JULIELab/EmoBank | |
- kingbri/PIPPA-shareGPT | |
license: apache-2.0 | |
quant of [jondurbin's](https://huggingface.co/jondurbin) [bagel-dpo-34b-v0.2](https://huggingface.co/jondurbin/bagel-dpo-34b-v0.2) | |
fits into 24gb with 16k context on windows | |
``` | |
python3 convert.py \ | |
-i /input/jondurbin_bagel-dpo-34b-v0.2/ \ | |
-c /input/pippa_cleaned/0000.parquet \ | |
-o /output/temp/ \ | |
-cf /output/bagel-dpo-34b-v0.2-4.65bpw-h6-exl2/ \ | |
-l 8192 \ | |
-ml 8192 \ | |
-b 4.65 \ | |
-hb 6 | |
``` |