chmjkb commited on
Commit
be5fb7c
·
1 Parent(s): 1c43a3c

Add config.json to each of the models

Browse files
llama-3.2-1B/QLoRA/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dim": 2048,
3
+ "n_layers": 16,
4
+ "n_heads": 32,
5
+ "n_kv_heads": 8,
6
+ "vocab_size": 128256,
7
+ "ffn_dim_multiplier": 1.5,
8
+ "multiple_of": 256,
9
+ "norm_eps": 1e-05,
10
+ "rope_theta": 500000.0,
11
+ "use_scaled_rope": true,
12
+ "quantization_args": {
13
+ "group_size": 32
14
+ },
15
+ "lora_args": {
16
+ "rank": 16,
17
+ "scale": 2.0
18
+ }
19
+ }
llama-3.2-1B/original/config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dim": 3072,
3
+ "n_layers": 28,
4
+ "n_heads": 24,
5
+ "n_kv_heads": 8,
6
+ "vocab_size": 128256,
7
+ "ffn_dim_multiplier": 1.0,
8
+ "multiple_of": 256,
9
+ "norm_eps": 1e-05,
10
+ "rope_theta": 500000.0,
11
+ "use_scaled_rope": true
12
+ }
llama-3.2-1B/spinquant/config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dim": 2048,
3
+ "n_layers": 16,
4
+ "n_heads": 32,
5
+ "n_kv_heads": 8,
6
+ "vocab_size": 128256,
7
+ "ffn_dim_multiplier": 1.5,
8
+ "multiple_of": 256,
9
+ "norm_eps": 1e-05,
10
+ "rope_theta": 500000.0,
11
+ "use_scaled_rope": true,
12
+ "quantization_args": {
13
+ "group_size": 32
14
+ }
15
+ }
llama-3.2-3B/QLoRA/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dim": 3072,
3
+ "n_layers": 28,
4
+ "n_heads": 24,
5
+ "n_kv_heads": 8,
6
+ "vocab_size": 128256,
7
+ "ffn_dim_multiplier": 1.0,
8
+ "multiple_of": 256,
9
+ "norm_eps": 1e-05,
10
+ "rope_theta": 500000.0,
11
+ "use_scaled_rope": true,
12
+ "quantization_args": {
13
+ "group_size": 32
14
+ },
15
+ "lora_args": {
16
+ "rank": 16,
17
+ "scale": 2.0
18
+ }
19
+ }
llama-3.2-3B/original/config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dim": 3072,
3
+ "n_layers": 28,
4
+ "n_heads": 24,
5
+ "n_kv_heads": 8,
6
+ "vocab_size": 128256,
7
+ "ffn_dim_multiplier": 1.0,
8
+ "multiple_of": 256,
9
+ "norm_eps": 1e-05,
10
+ "rope_theta": 500000.0,
11
+ "use_scaled_rope": true
12
+ }
llama-3.2-3B/spinquant/config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dim": 3072,
3
+ "ffn_dim_multiplier": 1.0,
4
+ "multiple_of": 256,
5
+ "n_heads": 24,
6
+ "n_kv_heads": 8,
7
+ "n_layers": 28,
8
+ "norm_eps": 1e-05,
9
+ "rope_theta": 500000.0,
10
+ "use_scaled_rope": true,
11
+ "vocab_size": 128256,
12
+ "quantization_args": {
13
+ "group_size": 32
14
+ }
15
+ }