Upload 7 files
Browse files- config.json +6 -5
- model.safetensors +2 -2
config.json
CHANGED
|
@@ -45,7 +45,7 @@
|
|
| 45 |
"audio_token_id": 258881,
|
| 46 |
"boa_token_id": 256000,
|
| 47 |
"boi_token_id": 255999,
|
| 48 |
-
"dtype": "
|
| 49 |
"eoa_token_id": 258883,
|
| 50 |
"eoa_token_index": 258883,
|
| 51 |
"eoi_token_id": 258882,
|
|
@@ -61,13 +61,13 @@
|
|
| 61 |
"attention_dropout": 0.0,
|
| 62 |
"attention_k_eq_v": false,
|
| 63 |
"bos_token_id": 2,
|
| 64 |
-
"dtype": "
|
| 65 |
"enable_moe_block": false,
|
| 66 |
"eos_token_id": 1,
|
| 67 |
"expert_intermediate_size": null,
|
| 68 |
"final_logit_softcapping": 30.0,
|
| 69 |
-
"global_head_dim":
|
| 70 |
-
"head_dim":
|
| 71 |
"hidden_activation": "gelu_pytorch_tanh",
|
| 72 |
"hidden_size": 8,
|
| 73 |
"hidden_size_per_layer_input": 1,
|
|
@@ -80,12 +80,13 @@
|
|
| 80 |
],
|
| 81 |
"max_position_embeddings": 131072,
|
| 82 |
"model_type": "gemma4_text",
|
|
|
|
| 83 |
"num_attention_heads": 2,
|
| 84 |
"num_experts": null,
|
| 85 |
"num_global_key_value_heads": null,
|
| 86 |
"num_hidden_layers": 3,
|
| 87 |
"num_key_value_heads": 1,
|
| 88 |
-
"num_kv_shared_layers":
|
| 89 |
"pad_token_id": 0,
|
| 90 |
"rms_norm_eps": 1e-06,
|
| 91 |
"rope_parameters": {
|
|
|
|
| 45 |
"audio_token_id": 258881,
|
| 46 |
"boa_token_id": 256000,
|
| 47 |
"boi_token_id": 255999,
|
| 48 |
+
"dtype": "float32",
|
| 49 |
"eoa_token_id": 258883,
|
| 50 |
"eoa_token_index": 258883,
|
| 51 |
"eoi_token_id": 258882,
|
|
|
|
| 61 |
"attention_dropout": 0.0,
|
| 62 |
"attention_k_eq_v": false,
|
| 63 |
"bos_token_id": 2,
|
| 64 |
+
"dtype": "float32",
|
| 65 |
"enable_moe_block": false,
|
| 66 |
"eos_token_id": 1,
|
| 67 |
"expert_intermediate_size": null,
|
| 68 |
"final_logit_softcapping": 30.0,
|
| 69 |
+
"global_head_dim": 4,
|
| 70 |
+
"head_dim": 4,
|
| 71 |
"hidden_activation": "gelu_pytorch_tanh",
|
| 72 |
"hidden_size": 8,
|
| 73 |
"hidden_size_per_layer_input": 1,
|
|
|
|
| 80 |
],
|
| 81 |
"max_position_embeddings": 131072,
|
| 82 |
"model_type": "gemma4_text",
|
| 83 |
+
"moe_intermediate_size": null,
|
| 84 |
"num_attention_heads": 2,
|
| 85 |
"num_experts": null,
|
| 86 |
"num_global_key_value_heads": null,
|
| 87 |
"num_hidden_layers": 3,
|
| 88 |
"num_key_value_heads": 1,
|
| 89 |
+
"num_kv_shared_layers": 1,
|
| 90 |
"pad_token_id": 0,
|
| 91 |
"rms_norm_eps": 1e-06,
|
| 92 |
"rope_parameters": {
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:feaf3cfe37e063524c5a46b94a145b50aeb7ea07d6b983070dda965ed0248c86
|
| 3 |
+
size 11997808
|