apepkuss79 commited on
Commit
f5d1d15
1 Parent(s): 78fdfd5

Update models

Browse files
.gitattributes CHANGED
@@ -33,3 +33,16 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ SmolLM2-1.7B-Instruct-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ SmolLM2-1.7B-Instruct-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ SmolLM2-1.7B-Instruct-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ SmolLM2-1.7B-Instruct-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ SmolLM2-1.7B-Instruct-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ SmolLM2-1.7B-Instruct-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ SmolLM2-1.7B-Instruct-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ SmolLM2-1.7B-Instruct-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ SmolLM2-1.7B-Instruct-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ SmolLM2-1.7B-Instruct-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ SmolLM2-1.7B-Instruct-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
47
+ SmolLM2-1.7B-Instruct-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
48
+ SmolLM2-1.7B-Instruct-f16.gguf filter=lfs diff=lfs merge=lfs -text
SmolLM2-1.7B-Instruct-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a81276a962be228e83df40e1be833b5735a902974e635d4931d6bc5d76053e0
3
+ size 674583392
SmolLM2-1.7B-Instruct-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:722e0f3846c0cd6613fc93580da4c1165de956d3626c55aed0b230a50194987f
3
+ size 932533088
SmolLM2-1.7B-Instruct-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2ce91737dfa2127c4fce6604b29218567f6343b16c560aab93efa025f958ecb
3
+ size 860181344
SmolLM2-1.7B-Instruct-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c34cbe3aeca752c09b39089a1c4fb521d7cb4699cc080cc6de9b03980c5b42d
3
+ size 776819552
SmolLM2-1.7B-Instruct-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d202c4c42d0bd299551cc099ac4fb2c12228b03f3bfa0d06cb00ab7f011d577
3
+ size 990729056
SmolLM2-1.7B-Instruct-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fab954d296aa25885960ed771a005343c220ac93ef78a53454e667e334935fe9
3
+ size 1055609696
SmolLM2-1.7B-Instruct-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b73b7bdaf7106c7a21da146b8ac4fd732afecffe591fb06f43313cdd57a4259
3
+ size 999117664
SmolLM2-1.7B-Instruct-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23ecf83012ddcbb17c8b9227b2cf7f6a4abff4b1ebfe2fec5caf9d7318243551
3
+ size 1192055648
SmolLM2-1.7B-Instruct-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57b99e35d3538d90385687402ab5cbaba76ba52618071f72fc467689b1ddaffa
3
+ size 1225479008
SmolLM2-1.7B-Instruct-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfe968c323792f71eff38719450aee55263bf7964bb9fddf61e33e9f2e540147
3
+ size 1192055648
SmolLM2-1.7B-Instruct-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a5e500d36887cb68be1b24ab7adefb07ea77b7d2e0d1000095025b5f7bc8a6f
3
+ size 1405965152
SmolLM2-1.7B-Instruct-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1812c04a6a47bf08da5d352d89978712db8225cb17d571c3c12805dd33411e6b
3
+ size 1820414816
SmolLM2-1.7B-Instruct-f16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84f3bea5ec6149793d9087205c433f11ac0de5602b942b59d5399773786a0b83
3
+ size 3424736096
config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LlamaForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 1,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 2048,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 8192,
13
+ "max_position_embeddings": 8192,
14
+ "mlp_bias": false,
15
+ "model_type": "llama",
16
+ "num_attention_heads": 32,
17
+ "num_hidden_layers": 24,
18
+ "num_key_value_heads": 32,
19
+ "pad_token_id": 2,
20
+ "pretraining_tp": 1,
21
+ "rms_norm_eps": 1e-05,
22
+ "rope_scaling": null,
23
+ "rope_theta": 130000,
24
+ "tie_word_embeddings": true,
25
+ "torch_dtype": "bfloat16",
26
+ "transformers_version": "4.42.3",
27
+ "transformers.js_config": {
28
+ "kv_cache_dtype": {
29
+ "q4f16": "float16",
30
+ "fp16": "float16"
31
+ }
32
+ },
33
+ "use_cache": true,
34
+ "vocab_size": 49152
35
+ }