Text Generation
Transformers
Safetensors
zamba2
Inference Endpoints
pglo commited on
Commit
6df099f
·
verified ·
1 Parent(s): e8e00bd

Upload config.json

Browse files
Files changed (1) hide show
  1. config.json +7 -1
config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "add_bias_linear": false,
3
  "architectures": [
4
  "Zamba2ForCausalLM"
@@ -69,8 +70,12 @@
69
  "m",
70
  "m"
71
  ],
 
 
72
  "lora_rank": 128,
73
  "mamba_headdim": 64,
 
 
74
  "model_type": "zamba2",
75
  "num_attention_heads": 32,
76
  "num_hidden_layers": 54,
@@ -86,5 +91,6 @@
86
  "use_cache": true,
87
  "use_mamba_kernels": true,
88
  "use_shared_block_lora": true,
89
- "vocab_size": 32000
 
90
  }
 
1
  {
2
+ "_name_or_path": "Zyphra/Zamba2-2.7B",
3
  "add_bias_linear": false,
4
  "architectures": [
5
  "Zamba2ForCausalLM"
 
70
  "m",
71
  "m"
72
  ],
73
+ "num_mem_blocks": 2,
74
+ "max_position_embeddings": 4096,
75
  "lora_rank": 128,
76
  "mamba_headdim": 64,
77
+ "use_mem_rope": false,
78
+ "rope_theta": 10000,
79
  "model_type": "zamba2",
80
  "num_attention_heads": 32,
81
  "num_hidden_layers": 54,
 
91
  "use_cache": true,
92
  "use_mamba_kernels": true,
93
  "use_shared_block_lora": true,
94
+ "se_shared_attention_lora": false,
95
+ "vocab_size": 32001
96
  }