MaziyarPanahi commited on
Commit
541783a
1 Parent(s): 477d899

Upload folder using huggingface_hub (#2)

Browse files

- 07b3cf58c1862764a90fecaf9f608b47723441706444ce23258a10c6355d86f9 (21a42a7438fca2c9b70dd672e4023e4ae9ee28ad)
- bd8bb1703c52fbc27b7e7b703f791bc91635a73eca1cfeb18125b64281de028d (63eb46ffd9dcc7748f2d8d08bd3ca32abb81ac7d)
- 0bc263b4215f034358c93324d4e34084ae32daff4b21fa6f096dc358867aed20 (ed7c04bb69f7f0961b14490c7f4b827ab55dadbb)
- fb75ce3d8c137bef8edecafd778ad754a21c0244a304825002b816f6002784f9 (e99fcdae07e9667cf4f34bed9849b07b6bec5a31)
- a7b55e4dd2c9dfddc5ba2dbc9dc1c9e8691b02d6f81bf75d03d3d5cfac66f9bf (cc8c835188b6a99054404d587477267468c1ffa8)
- 1ce5035b135261e8d2a4fd708a47ca66266de3131e324c582d3a932df9be29b2 (1169a655225734e39ae096d7024acae57dfc56bd)
- 994c93e70ebc319cbef858d856ea23b083b8be32e2d8dd9628c91aa79df7e432 (b6e6b4d58a981876abdbb119b096f860d818238a)
- c7fec94d7fe1386d8a03b1df4b93967f165e3bea0ca5249909523faa3fc696bf (2bddbcbe199a6daa92d9459dd5b90f6e14f12617)
- 63770611d12348b12abd16a94b63c03ed7220e717659e4f7ee51cb973e7dc81f (a7f36682f8e423b5459ea97226ea6367364c9293)
- e35207287542a3e89b8c476e0e7a51a2434a7a9913b099566104d236fc1b116c (2e178bc4e3a2130222f98aab6f813fe8e828e147)
- eeecca46d478625dac5d17574120522c6975e18eb8269eaf8bc9dbcd8113de28 (77619691f0148ab619362b462c9a7524a263887e)
- a73aef439e689e5bfd233dcb4c959c1d03f7b69f7d0be38950c8f526d4bbea61 (ab7db8d0ec76c2a773d9b5c5f16a4c89a9d88417)
- 38d3509ce766513f2f506d576c12c8ee6150c9208cc80644b969f9b9a2f7d5d5 (8c6897e4c1f63387b3a5707291debb78da4060e6)

.gitattributes CHANGED
@@ -43,3 +43,5 @@ tulu-2-dpo-13b.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
43
  tulu-2-dpo-13b.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
44
  tulu-2-dpo-13b.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
45
  tulu-2-dpo-13b.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
 
 
 
43
  tulu-2-dpo-13b.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
44
  tulu-2-dpo-13b.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
45
  tulu-2-dpo-13b.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
46
+ tulu-2-dpo-13b.IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text
47
+ tulu-2-dpo-13b.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -10,6 +10,7 @@ tags:
10
  - GGUF
11
  - transformers
12
  - pytorch
 
13
  - llama
14
  - text-generation
15
  - en
@@ -98,7 +99,7 @@ pip3 install huggingface-hub
98
  Then you can download any individual model file to the current directory, at high speed, with a command like this:
99
 
100
  ```shell
101
- huggingface-cli download MaziyarPanahi/tulu-2-dpo-13b-GGUF tulu-2-dpo-13b-GGUF.Q4_K_M.gguf --local-dir . --local-dir-use-symlinks False
102
  ```
103
  </details>
104
  <details>
@@ -121,7 +122,7 @@ pip3 install hf_transfer
121
  And set environment variable `HF_HUB_ENABLE_HF_TRANSFER` to `1`:
122
 
123
  ```shell
124
- HF_HUB_ENABLE_HF_TRANSFER=1 huggingface-cli download MaziyarPanahi/tulu-2-dpo-13b-GGUF tulu-2-dpo-13b-GGUF.Q4_K_M.gguf --local-dir . --local-dir-use-symlinks False
125
  ```
126
 
127
  Windows Command Line users: You can set the environment variable by running `set HF_HUB_ENABLE_HF_TRANSFER=1` before the download command.
@@ -132,7 +133,7 @@ Windows Command Line users: You can set the environment variable by running `set
132
  Make sure you are using `llama.cpp` from commit [d0cee0d](https://github.com/ggerganov/llama.cpp/commit/d0cee0d36d5be95a0d9088b674dbb27354107221) or later.
133
 
134
  ```shell
135
- ./main -ngl 35 -m tulu-2-dpo-13b-GGUF.Q4_K_M.gguf --color -c 32768 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "<|im_start|>system
136
  {system_message}<|im_end|>
137
  <|im_start|>user
138
  {prompt}<|im_end|>
@@ -149,7 +150,7 @@ For other parameters and how to use them, please refer to [the llama.cpp documen
149
 
150
  ## How to run in `text-generation-webui`
151
 
152
- Further instructions can be found in the text-generation-webui documentation, here: [text-generation-webui/docs/04 ‐ Model Tab.md](https://github.com/oobabooga/text-generation-webui/blob/main/docs/04%20%E2%80%90%20Model%20Tab.md#llamacpp).
153
 
154
  ## How to run from Python code
155
 
@@ -157,7 +158,7 @@ You can use GGUF models from Python using the [llama-cpp-python](https://github.
157
 
158
  ### How to load this model in Python code, using llama-cpp-python
159
 
160
- For full documentation, please see: [llama-cpp-python docs](https://abetlen.github.io/llama-cpp-python/).
161
 
162
  #### First install the package
163
 
@@ -189,7 +190,7 @@ from llama_cpp import Llama
189
 
190
  # Set gpu_layers to the number of layers to offload to GPU. Set to 0 if no GPU acceleration is available on your system.
191
  llm = Llama(
192
- model_path="./tulu-2-dpo-13b-GGUF.Q4_K_M.gguf", # Download the model file first
193
  n_ctx=32768, # The max sequence length to use - note that longer sequence lengths require much more resources
194
  n_threads=8, # The number of CPU threads to use, tailor to your system and the resulting performance
195
  n_gpu_layers=35 # The number of layers to offload to GPU, if you have GPU acceleration available
@@ -209,7 +210,7 @@ output = llm(
209
 
210
  # Chat Completion API
211
 
212
- llm = Llama(model_path="./tulu-2-dpo-13b-GGUF.Q4_K_M.gguf", chat_format="llama-2") # Set chat_format according to the model you are using
213
  llm.create_chat_completion(
214
  messages = [
215
  {"role": "system", "content": "You are a story writing assistant."},
 
10
  - GGUF
11
  - transformers
12
  - pytorch
13
+ - safetensors
14
  - llama
15
  - text-generation
16
  - en
 
99
  Then you can download any individual model file to the current directory, at high speed, with a command like this:
100
 
101
  ```shell
102
+ huggingface-cli download MaziyarPanahi/tulu-2-dpo-13b-GGUF tulu-2-dpo-13b.Q4_K_M.gguf --local-dir . --local-dir-use-symlinks False
103
  ```
104
  </details>
105
  <details>
 
122
  And set environment variable `HF_HUB_ENABLE_HF_TRANSFER` to `1`:
123
 
124
  ```shell
125
+ HF_HUB_ENABLE_HF_TRANSFER=1 huggingface-cli download MaziyarPanahi/tulu-2-dpo-13b-GGUF tulu-2-dpo-13b.Q4_K_M.gguf --local-dir . --local-dir-use-symlinks False
126
  ```
127
 
128
  Windows Command Line users: You can set the environment variable by running `set HF_HUB_ENABLE_HF_TRANSFER=1` before the download command.
 
133
  Make sure you are using `llama.cpp` from commit [d0cee0d](https://github.com/ggerganov/llama.cpp/commit/d0cee0d36d5be95a0d9088b674dbb27354107221) or later.
134
 
135
  ```shell
136
+ ./main -ngl 35 -m tulu-2-dpo-13b.Q4_K_M.gguf --color -c 32768 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "<|im_start|>system
137
  {system_message}<|im_end|>
138
  <|im_start|>user
139
  {prompt}<|im_end|>
 
150
 
151
  ## How to run in `text-generation-webui`
152
 
153
+ Further instructions can be found in the text-generation-webui documentation, here: [text-generation-webui/docs/04 ‐ Model Tab.md](https://github.com/oobabooga/text-generation-webui/blob/main/docs/04%20-%20Model%20Tab.md#llamacpp).
154
 
155
  ## How to run from Python code
156
 
 
158
 
159
  ### How to load this model in Python code, using llama-cpp-python
160
 
161
+ For full documentation, please see: [llama-cpp-python docs](https://github.com/abetlen/llama-cpp-python/).
162
 
163
  #### First install the package
164
 
 
190
 
191
  # Set gpu_layers to the number of layers to offload to GPU. Set to 0 if no GPU acceleration is available on your system.
192
  llm = Llama(
193
+ model_path="./tulu-2-dpo-13b.Q4_K_M.gguf", # Download the model file first
194
  n_ctx=32768, # The max sequence length to use - note that longer sequence lengths require much more resources
195
  n_threads=8, # The number of CPU threads to use, tailor to your system and the resulting performance
196
  n_gpu_layers=35 # The number of layers to offload to GPU, if you have GPU acceleration available
 
210
 
211
  # Chat Completion API
212
 
213
+ llm = Llama(model_path="./tulu-2-dpo-13b.Q4_K_M.gguf", chat_format="llama-2") # Set chat_format according to the model you are using
214
  llm.create_chat_completion(
215
  messages = [
216
  {"role": "system", "content": "You are a story writing assistant."},
tulu-2-dpo-13b.IQ3_XS.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a4179d21f4435c64d8fff0b7bd65cb1086169c10f7c84d2a6daccf7adfbaf8c
3
+ size 5361610944
tulu-2-dpo-13b.IQ4_XS.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffbee25bd18bda9bf6f409ad876fae1bd30de5b53c3d7764978fd02554fae058
3
+ size 7019518144
tulu-2-dpo-13b.Q2_K.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:31403fb7dc715913bd9cbefe668db492524e847f4fc55eb44f48abea3d449066
3
- size 4854269920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ae74fb7d1bc709a348b650134f50a66e541d64c31006f526b106733016857ee
3
+ size 4854270144
tulu-2-dpo-13b.Q3_K_L.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6a8ef92e9f31add4a59433143cc57f8c7d7140a674f5592e27aa1ffe9734ce67
3
- size 6929559520
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0fc9d4b2846fdff1d2a0a39258cd1572ce51ee5ae0ff17b9185975cbfbc1326
3
+ size 6929559744
tulu-2-dpo-13b.Q3_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:df633e14ac92240a903566c510eaaddc13ab13a8a1885c0c1c8af68c24b70ad4
3
- size 6337769440
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86977b49e1cd0b5b7dcc8fbfbb3667ad7d39b671865e8adb37c3216a73518fc8
3
+ size 6337769664
tulu-2-dpo-13b.Q3_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b6da25c2b413c750e1aceb0df528171315d67c2baff8a62d129e6ee28913fcc8
3
- size 5658980320
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8b6b66fc333972aa64380427d5e3362fbe91e3588d2827c7856ee8dc0e8f5a3
3
+ size 5658980544
tulu-2-dpo-13b.Q4_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ef4d4b25adb5c74648fcf19bbbd0adc753e58add7ce05bda9ec11fc22c7f1a5c
3
- size 7865956320
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59e253478d970b952d8204d0ad52c3f87959d027f7b240c1c00ba5107e8d36d5
3
+ size 7865956544
tulu-2-dpo-13b.Q4_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:88247de4d2832d16463285c14851a6c688406b6892b8533a8d78e7a9b13a448c
3
- size 7423178720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53347c88ca23c6f636421f45d538f7a7e47a1b8be10966b01fcbb488b37c1fb3
3
+ size 7423178944
tulu-2-dpo-13b.Q5_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b8003c551efebdc2ed458b90a7f2e7b0e171d7c25d88cd20d1148843d163f5c2
3
- size 9229924320
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1dd54cb5f6b88b72dba5348c5d1da3ce20b140b380fa3d84aaf300cd9031b213
3
+ size 9229924544
tulu-2-dpo-13b.Q5_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:44e702bcdeb37e157e9ea35ac4e938d7d2f1b0d7319aeec75c655790a0ad9281
3
- size 8972285920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d7b4264fa1e76563274dc6936dc3e2dcaa600eca10c9bb1f2145e5eb8cd9570
3
+ size 8972286144
tulu-2-dpo-13b.Q6_K.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7993c2a52aba720723fb0c60110c22251df2aa4f78e9e62645a347785e0bb463
3
- size 10679140320
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70d83833cf611dbdbb423b605a3381b91129a519cc2fbbf0b13185b898c73716
3
+ size 10679140544
tulu-2-dpo-13b.Q8_0.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a1e7a1c5d597d8a59cb8eb5c212dfce8d839b25fdf4b524ad7a62bdbf72f3db
3
- size 13831319520
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6fbb27802565c81c9b18678359e1cf6a930601e1d3ab0a33607606aa4ef99b07
3
+ size 13831319744