Suparious commited on
Commit
6f9baf4
1 Parent(s): 0b19469

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +116 -0
README.md ADDED
@@ -0,0 +1,116 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: yi-license
4
+ license_link: https://huggingface.co/01-ai/Yi-34B-200K/blob/main/LICENSE
5
+ base_model: 01-ai/yi-34b-200k
6
+ datasets:
7
+ - ai2_arc
8
+ - allenai/ultrafeedback_binarized_cleaned
9
+ - argilla/distilabel-intel-orca-dpo-pairs
10
+ - jondurbin/airoboros-3.2
11
+ - codeparrot/apps
12
+ - facebook/belebele
13
+ - bluemoon-fandom-1-1-rp-cleaned
14
+ - boolq
15
+ - camel-ai/biology
16
+ - camel-ai/chemistry
17
+ - camel-ai/math
18
+ - camel-ai/physics
19
+ - jondurbin/contextual-dpo-v0.1
20
+ - jondurbin/gutenberg-dpo-v0.1
21
+ - jondurbin/py-dpo-v0.1
22
+ - jondurbin/truthy-dpo-v0.1
23
+ - LDJnr/Capybara
24
+ - jondurbin/cinematika-v0.1
25
+ - WizardLM/WizardLM_evol_instruct_70k
26
+ - glaiveai/glaive-function-calling-v2
27
+ - jondurbin/gutenberg-dpo-v0.1
28
+ - grimulkan/LimaRP-augmented
29
+ - lmsys/lmsys-chat-1m
30
+ - ParisNeo/lollms_aware_dataset
31
+ - TIGER-Lab/MathInstruct
32
+ - Muennighoff/natural-instructions
33
+ - openbookqa
34
+ - kingbri/PIPPA-shareGPT
35
+ - piqa
36
+ - Vezora/Tested-22k-Python-Alpaca
37
+ - ropes
38
+ - cakiki/rosetta-code
39
+ - Open-Orca/SlimOrca
40
+ - b-mc2/sql-create-context
41
+ - squad_v2
42
+ - mattpscott/airoboros-summarization
43
+ - migtissera/Synthia-v1.3
44
+ - unalignment/toxic-dpo-v0.2
45
+ - WhiteRabbitNeo/WRN-Chapter-1
46
+ - WhiteRabbitNeo/WRN-Chapter-2
47
+ - winogrande
48
+ ---
49
+ # jondurbin/bagel-34b-v0.5 Exl2
50
+
51
+ - Model creator: [jondurbin](https://huggingface.co/jondurbin)
52
+ - Original model: [bagel-34b-v0.5](https://huggingface.co/jondurbin/bagel-34b-v0.5)
53
+
54
+ ![bagel](bagel.png)
55
+
56
+ ## Model Summary
57
+
58
+ This is a fine-tune of the updated yi-34b-200k with better long-context support.
59
+
60
+ See [bagel](https://github.com/jondurbin/bagel) for additional details on the datasets.
61
+
62
+ The DPO version is available [here](https://huggingface.co/jondurbin/bagel-dpo-34b-v0.5)
63
+
64
+ ## How to Use
65
+
66
+ Using <a href="https://github.com/turboderp/exllamav2/releases/tag/v0.0.14">turboderp's ExLlamaV2 v0.0.14</a> for quantization.
67
+
68
+ <b>The "main" branch only contains the measurement.json, download one of the other branches for the model (see below)</b>
69
+
70
+ Each branch contains an individual bits per weight, with the main one containing only the meaurement.json for further conversions.
71
+
72
+ Original model: https://huggingface.co/jondurbin/bagel-34b-v0.5
73
+
74
+ | Branch | Bits | lm_head bits | VRAM (4k) | VRAM (16k) | VRAM (32k) | Description |
75
+ | ------ | ---- | ------------ | ---- | ---- | ---- | ----------- |
76
+ | [6_5](https://huggingface.co/suparious/bagel-34b-v0.5-exl2/tree/6_5) | 6.5 | 8.0 | 28.9 GB | 31.6 GB | 35.6 GB | Near unquantized performance at vastly reduced size, **recommended**. |
77
+ | [4_25](https://huggingface.co/suparious/bagel-34b-v0.5-exl2/tree/4_25) | 4.25 | 6.0 | 19.5 GB | 22.2 GB | 26.2 GB | GPTQ equivalent bits per weight, slightly higher quality. |
78
+ | [3_5](https://huggingface.co/suparious/bagel-34b-v0.5-exl2/tree/3_5) | 3.5 | 6.0 | 16.5 GB | 19.2 GB | 23.2 GB | Lower quality, only use if you have to. |
79
+ | [3_0](https://huggingface.co/suparious/bagel-34b-v0.5-exl2/tree/3_0) | 3.0 | 6.0 | 14.3 GB | 17.0 GB | 21.0 GB | Very low quality, usable with 16gb of VRAM. |
80
+
81
+ ## Download instructions
82
+
83
+ With git:
84
+
85
+ ```shell
86
+ git clone --single-branch --branch 6_5 https://huggingface.co/suparious/bagel-34b-v0.5-exl2 bagel-34b-v0.5-exl2-6_5
87
+ ```
88
+
89
+ With huggingface hub (credit to TheBloke for instructions):
90
+
91
+ ```shell
92
+ pip3 install huggingface-hub
93
+ ```
94
+
95
+ To download the `main` (only useful if you only care about measurement.json) branch to a folder called `bagel-34b-v0.5-exl2`:
96
+
97
+ ```shell
98
+ mkdir bagel-34b-v0.5-exl2
99
+ huggingface-cli download suparious/bagel-34b-v0.5-exl2 --local-dir bagel-34b-v0.5-exl2 --local-dir-use-symlinks False
100
+ ```
101
+
102
+ To download from a different branch, add the `--revision` parameter:
103
+
104
+ Linux:
105
+
106
+ ```shell
107
+ mkdir bagel-34b-v0.5-exl2-6_5
108
+ huggingface-cli download suparious/bagel-34b-v0.5-exl2 --revision 6_5 --local-dir bagel-34b-v0.5-exl2-6_5 --local-dir-use-symlinks False
109
+ ```
110
+
111
+ Windows (which apparently doesn't like _ in folders sometimes?):
112
+
113
+ ```shell
114
+ mkdir bagel-34b-v0.5-exl2-6.5
115
+ huggingface-cli download suparious/bagel-34b-v0.5-exl2 --revision 6_5 --local-dir bagel-34b-v0.5-exl2-6.5 --local-dir-use-symlinks False
116
+ ```