Suparious commited on
Commit
eff9fdb
·
verified ·
1 Parent(s): 4b34f25

Add model card

Browse files
Files changed (1) hide show
  1. README.md +122 -0
README.md ADDED
@@ -0,0 +1,122 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: yi-license
4
+ license_link: https://huggingface.co/01-ai/Yi-34B-200K/blob/main/LICENSE
5
+ base_model: 01-ai/yi-34b-200k
6
+ datasets:
7
+ - ai2_arc
8
+ - allenai/ultrafeedback_binarized_cleaned
9
+ - argilla/distilabel-intel-orca-dpo-pairs
10
+ - jondurbin/airoboros-3.2
11
+ - codeparrot/apps
12
+ - facebook/belebele
13
+ - bluemoon-fandom-1-1-rp-cleaned
14
+ - boolq
15
+ - camel-ai/biology
16
+ - camel-ai/chemistry
17
+ - camel-ai/math
18
+ - camel-ai/physics
19
+ - jondurbin/contextual-dpo-v0.1
20
+ - jondurbin/gutenberg-dpo-v0.1
21
+ - jondurbin/py-dpo-v0.1
22
+ - jondurbin/truthy-dpo-v0.1
23
+ - LDJnr/Capybara
24
+ - jondurbin/cinematika-v0.1
25
+ - WizardLM/WizardLM_evol_instruct_70k
26
+ - glaiveai/glaive-function-calling-v2
27
+ - jondurbin/gutenberg-dpo-v0.1
28
+ - grimulkan/LimaRP-augmented
29
+ - lmsys/lmsys-chat-1m
30
+ - ParisNeo/lollms_aware_dataset
31
+ - TIGER-Lab/MathInstruct
32
+ - Muennighoff/natural-instructions
33
+ - openbookqa
34
+ - kingbri/PIPPA-shareGPT
35
+ - piqa
36
+ - Vezora/Tested-22k-Python-Alpaca
37
+ - ropes
38
+ - cakiki/rosetta-code
39
+ - Open-Orca/SlimOrca
40
+ - b-mc2/sql-create-context
41
+ - squad_v2
42
+ - mattpscott/airoboros-summarization
43
+ - migtissera/Synthia-v1.3
44
+ - unalignment/toxic-dpo-v0.2
45
+ - WhiteRabbitNeo/WRN-Chapter-1
46
+ - WhiteRabbitNeo/WRN-Chapter-2
47
+ - winogrande
48
+ tags:
49
+ - quantized
50
+ - chatml
51
+ model_type: mistral
52
+ pipeline_tag: text-generation
53
+ inference: false
54
+ ---
55
+ # jondurbin/bagel-dpo-34b-v0.5 Exl2
56
+
57
+ - Model creator: [jondurbin](https://huggingface.co/jondurbin)
58
+ - Original model: [bagel-dpo-34b-v0.5](https://huggingface.co/jondurbin/bagel-dpo-34b-v0.5)
59
+
60
+ ![bagel](bagel.png)
61
+
62
+ ## Model Summary
63
+
64
+ This is a fine-tune of the updated yi-34b-200k with better long-context support.
65
+
66
+ See [bagel](https://github.com/jondurbin/bagel) for additional details on the datasets.
67
+
68
+ The DPO version is available [here](https://huggingface.co/jondurbin/bagel-dpo-34b-v0.5)
69
+
70
+ ## How to Use
71
+
72
+ Using <a href="https://github.com/turboderp/exllamav2/releases/tag/v0.0.14">turboderp's ExLlamaV2 v0.0.14</a> for quantization.
73
+
74
+ <b>The "main" branch only contains the measurement.json, download one of the other branches for the model (see below)</b>
75
+
76
+ Each branch contains an individual bits per weight, with the main one containing only the meaurement.json for further conversions.
77
+
78
+ Original model: https://huggingface.co/jondurbin/bagel-dpo-34b-v0.5
79
+
80
+ | Branch | Bits | lm_head bits | VRAM (4k) | VRAM (16k) | VRAM (32k) | Description |
81
+ | ------ | ---- | ------------ | ---- | ---- | ---- | ----------- |
82
+ | [6_5](https://huggingface.co/suparious/bagel-dpo-34b-v0.5-exl2/tree/6_5) | 6.5 | 8.0 | 28.9 GB | 31.6 GB | 35.6 GB | Near unquantized performance at vastly reduced size, **recommended**. |
83
+ | [4_25](https://huggingface.co/suparious/bagel-dpo-34b-v0.5-exl2/tree/4_25) | 4.25 | 6.0 | 19.5 GB | 22.2 GB | 26.2 GB | GPTQ equivalent bits per weight, slightly higher quality. |
84
+ | [3_5](https://huggingface.co/suparious/bagel-dpo-34b-v0.5-exl2/tree/3_5) | 3.5 | 6.0 | 16.5 GB | 19.2 GB | 23.2 GB | Lower quality, only use if you have to. |
85
+ | [3_0](https://huggingface.co/suparious/bagel-dpo-34b-v0.5-exl2/tree/3_0) | 3.0 | 6.0 | 14.3 GB | 17.0 GB | 21.0 GB | Very low quality, usable with 16gb of VRAM. |
86
+
87
+ ## Download instructions
88
+
89
+ With git:
90
+
91
+ ```shell
92
+ git clone --single-branch --branch 6_5 https://huggingface.co/suparious/bagel-dpo-34b-v0.5-exl2 bagel-dpo-34b-v0.5-exl2-6_5
93
+ ```
94
+
95
+ With huggingface hub (credit to TheBloke for instructions):
96
+
97
+ ```shell
98
+ pip3 install huggingface-hub
99
+ ```
100
+
101
+ To download the `main` (only useful if you only care about measurement.json) branch to a folder called `bagel-dpo-34b-v0.5-exl2`:
102
+
103
+ ```shell
104
+ mkdir bagel-dpo-34b-v0.5-exl2
105
+ huggingface-cli download suparious/bagel-dpo-34b-v0.5-exl2 --local-dir bagel-dpo-34b-v0.5-exl2 --local-dir-use-symlinks False
106
+ ```
107
+
108
+ To download from a different branch, add the `--revision` parameter:
109
+
110
+ Linux:
111
+
112
+ ```shell
113
+ mkdir bagel-dpo-34b-v0.5-exl2-6_5
114
+ huggingface-cli download suparious/bagel-dpo-34b-v0.5-exl2 --revision 6_5 --local-dir bagel-dpo-34b-v0.5-exl2-6_5 --local-dir-use-symlinks False
115
+ ```
116
+
117
+ Windows (which apparently doesn't like _ in folders sometimes?):
118
+
119
+ ```shell
120
+ mkdir bagel-dpo-34b-v0.5-exl2-6.5
121
+ huggingface-cli download suparious/bagel-dpo-34b-v0.5-exl2 --revision 6_5 --local-dir bagel-dpo-34b-v0.5-exl2-6.5 --local-dir-use-symlinks False
122
+ ```