Update models
Browse files- .gitattributes +15 -0
- QwQ-32B-Q2_K.gguf +3 -0
- QwQ-32B-Q3_K_L.gguf +3 -0
- QwQ-32B-Q3_K_M.gguf +3 -0
- QwQ-32B-Q3_K_S.gguf +3 -0
- QwQ-32B-Q4_0.gguf +3 -0
- QwQ-32B-Q4_K_M.gguf +3 -0
- QwQ-32B-Q4_K_S.gguf +3 -0
- QwQ-32B-Q5_0.gguf +3 -0
- QwQ-32B-Q5_K_M.gguf +3 -0
- QwQ-32B-Q5_K_S.gguf +3 -0
- QwQ-32B-Q6_K.gguf +3 -0
- QwQ-32B-Q8_0.gguf +3 -0
- QwQ-32B-f16-00001-of-00003.gguf +3 -0
- QwQ-32B-f16-00002-of-00003.gguf +3 -0
- QwQ-32B-f16-00003-of-00003.gguf +3 -0
- config.json +27 -0
.gitattributes
CHANGED
@@ -33,3 +33,18 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
QwQ-32B-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
37 |
+
QwQ-32B-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
|
38 |
+
QwQ-32B-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
39 |
+
QwQ-32B-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
40 |
+
QwQ-32B-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
|
41 |
+
QwQ-32B-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
42 |
+
QwQ-32B-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
43 |
+
QwQ-32B-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
|
44 |
+
QwQ-32B-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
45 |
+
QwQ-32B-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
46 |
+
QwQ-32B-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
47 |
+
QwQ-32B-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
48 |
+
QwQ-32B-f16-00001-of-00003.gguf filter=lfs diff=lfs merge=lfs -text
|
49 |
+
QwQ-32B-f16-00002-of-00003.gguf filter=lfs diff=lfs merge=lfs -text
|
50 |
+
QwQ-32B-f16-00003-of-00003.gguf filter=lfs diff=lfs merge=lfs -text
|
QwQ-32B-Q2_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:89ee455ecbbb0d9395b6d60dddbeaedbf36b25452d8774dfea09ca3beb23808a
|
3 |
+
size 12313098816
|
QwQ-32B-Q3_K_L.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9207f1609a6d3cd0e972a9c708dea748f36a5c06e3f5d47b997aed3c8098bdb5
|
3 |
+
size 17247078976
|
QwQ-32B-Q3_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cf5812b92ce09e8d58c6f1298207daabcaba0f106bb3358d50da3573d31e7acc
|
3 |
+
size 15935048256
|
QwQ-32B-Q3_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9ad80c156f9b388ec1f50a51ffe78c7768262bdfc735d509a76dd67fd7342bf1
|
3 |
+
size 14392330816
|
QwQ-32B-Q4_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:794244be231887f9fd04eeae23ba5a995936d7b5d4fe53b59667a42ec46ef3ac
|
3 |
+
size 18640230976
|
QwQ-32B-Q4_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0e9b6cb902cc17dbd5ffbc481dc9303cd2ea24daf3b49398d474bfeb22e419e4
|
3 |
+
size 19851336256
|
QwQ-32B-Q4_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9d258617bdf056d583c0723a43abc7f8cb8b6ae8d80e3024daa373b58bc43730
|
3 |
+
size 18784410176
|
QwQ-32B-Q5_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f5a1c01c35d96351fe08877eb942ce4e66b903c458680905aa069f9b010ec3f0
|
3 |
+
size 22638254656
|
QwQ-32B-Q5_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c65883d252c5707ed98b05f980515963f53f95bdb5caa644098be1be000cba51
|
3 |
+
size 23262157376
|
QwQ-32B-Q5_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ee94b96eb0b8a6f01600185f8fbcdeeae5d9dcc874289c13608558a0eeeff9bc
|
3 |
+
size 22638254656
|
QwQ-32B-Q6_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:13f9fda244c5a73bfb6464062a4666c715c18834d4ec438533f9c8d884925f53
|
3 |
+
size 26886154816
|
QwQ-32B-Q8_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6c0b473616167f14af406d122ed3a12f374f6958d2ed2a5af0a889e0cf9f1695
|
3 |
+
size 34820885056
|
QwQ-32B-f16-00001-of-00003.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:db91081a5379f6c92aa80b4e34560ef9cb9bc12e47621ca964799a72e9c45715
|
3 |
+
size 29845224544
|
QwQ-32B-f16-00002-of-00003.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a060881be66e15f9f01ae910c26103ae52a40c748d03623d73ca8e95f38c68b6
|
3 |
+
size 29823611296
|
QwQ-32B-f16-00003-of-00003.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7b7e242c5e78dc962f4acb0ed7fea1a109d4f798ceeed4d589d7e0ed091edba0
|
3 |
+
size 5867134368
|
config.json
ADDED
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"Qwen2ForCausalLM"
|
4 |
+
],
|
5 |
+
"attention_dropout": 0.0,
|
6 |
+
"bos_token_id": 151643,
|
7 |
+
"eos_token_id": 151645,
|
8 |
+
"hidden_act": "silu",
|
9 |
+
"hidden_size": 5120,
|
10 |
+
"initializer_range": 0.02,
|
11 |
+
"intermediate_size": 27648,
|
12 |
+
"max_position_embeddings": 131072,
|
13 |
+
"max_window_layers": 64,
|
14 |
+
"model_type": "qwen2",
|
15 |
+
"num_attention_heads": 40,
|
16 |
+
"num_hidden_layers": 64,
|
17 |
+
"num_key_value_heads": 8,
|
18 |
+
"rms_norm_eps": 1e-05,
|
19 |
+
"rope_theta": 1000000.0,
|
20 |
+
"sliding_window": 32768,
|
21 |
+
"tie_word_embeddings": false,
|
22 |
+
"torch_dtype": "bfloat16",
|
23 |
+
"transformers_version": "4.43.1",
|
24 |
+
"use_cache": true,
|
25 |
+
"use_sliding_window": false,
|
26 |
+
"vocab_size": 152064
|
27 |
+
}
|