rwkv6-7B-finch / model.safetensors.index.json
yzhangcs's picture
Upload RWKV6ForCausalLM
44f743b verified
raw
history blame
61.7 kB
{
"metadata": {
"total_size": 15271493632
},
"weight_map": {
"lm_head.weight": "model-00004-of-00004.safetensors",
"model.embeddings.weight": "model-00001-of-00004.safetensors",
"model.layers.0.attn.bonus": "model-00001-of-00004.safetensors",
"model.layers.0.attn.g_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.0.attn.g_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.0.attn.g_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.0.attn.k_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.0.attn.o_proj.weight": "model-00001-of-00004.safetensors",
"model.layers.0.attn.r_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.0.attn.v_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.0.attn.w_proj.linear.lora.0.weight": "model-00001-of-00004.safetensors",
"model.layers.0.attn.w_proj.linear.lora.2.bias": "model-00001-of-00004.safetensors",
"model.layers.0.attn.w_proj.linear.lora.2.weight": "model-00001-of-00004.safetensors",
"model.layers.0.attn.x_bias": "model-00001-of-00004.safetensors",
"model.layers.0.attn.x_proj.0.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.0.attn.x_proj.0.mu": "model-00001-of-00004.safetensors",
"model.layers.0.attn.x_proj.2.weight": "model-00001-of-00004.safetensors",
"model.layers.0.attn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.0.attn_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.0.ffn.key.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.0.ffn.key.mu": "model-00001-of-00004.safetensors",
"model.layers.0.ffn.receptance.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.0.ffn.receptance.mu": "model-00001-of-00004.safetensors",
"model.layers.0.ffn.value.weight": "model-00001-of-00004.safetensors",
"model.layers.0.ffn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.0.ffn_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.0.pre_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.0.pre_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.1.attn.bonus": "model-00001-of-00004.safetensors",
"model.layers.1.attn.g_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.1.attn.g_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.1.attn.g_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.1.attn.k_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.1.attn.o_proj.weight": "model-00001-of-00004.safetensors",
"model.layers.1.attn.r_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.1.attn.v_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.1.attn.w_proj.linear.lora.0.weight": "model-00001-of-00004.safetensors",
"model.layers.1.attn.w_proj.linear.lora.2.bias": "model-00001-of-00004.safetensors",
"model.layers.1.attn.w_proj.linear.lora.2.weight": "model-00001-of-00004.safetensors",
"model.layers.1.attn.x_bias": "model-00001-of-00004.safetensors",
"model.layers.1.attn.x_proj.0.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.1.attn.x_proj.0.mu": "model-00001-of-00004.safetensors",
"model.layers.1.attn.x_proj.2.weight": "model-00001-of-00004.safetensors",
"model.layers.1.attn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.1.attn_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.1.ffn.key.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.1.ffn.key.mu": "model-00001-of-00004.safetensors",
"model.layers.1.ffn.receptance.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.1.ffn.receptance.mu": "model-00001-of-00004.safetensors",
"model.layers.1.ffn.value.weight": "model-00001-of-00004.safetensors",
"model.layers.1.ffn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.1.ffn_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.10.attn.bonus": "model-00001-of-00004.safetensors",
"model.layers.10.attn.g_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.10.attn.g_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.10.attn.g_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.10.attn.k_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.10.attn.o_proj.weight": "model-00002-of-00004.safetensors",
"model.layers.10.attn.r_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.10.attn.v_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.10.attn.w_proj.linear.lora.0.weight": "model-00002-of-00004.safetensors",
"model.layers.10.attn.w_proj.linear.lora.2.bias": "model-00002-of-00004.safetensors",
"model.layers.10.attn.w_proj.linear.lora.2.weight": "model-00002-of-00004.safetensors",
"model.layers.10.attn.x_bias": "model-00001-of-00004.safetensors",
"model.layers.10.attn.x_proj.0.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.10.attn.x_proj.0.mu": "model-00001-of-00004.safetensors",
"model.layers.10.attn.x_proj.2.weight": "model-00001-of-00004.safetensors",
"model.layers.10.attn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.10.attn_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.10.ffn.key.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.10.ffn.key.mu": "model-00002-of-00004.safetensors",
"model.layers.10.ffn.receptance.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.10.ffn.receptance.mu": "model-00002-of-00004.safetensors",
"model.layers.10.ffn.value.weight": "model-00002-of-00004.safetensors",
"model.layers.10.ffn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.10.ffn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.11.attn.bonus": "model-00002-of-00004.safetensors",
"model.layers.11.attn.g_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.11.attn.g_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.11.attn.g_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.11.attn.k_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.11.attn.o_proj.weight": "model-00002-of-00004.safetensors",
"model.layers.11.attn.r_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.11.attn.v_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.11.attn.w_proj.linear.lora.0.weight": "model-00002-of-00004.safetensors",
"model.layers.11.attn.w_proj.linear.lora.2.bias": "model-00002-of-00004.safetensors",
"model.layers.11.attn.w_proj.linear.lora.2.weight": "model-00002-of-00004.safetensors",
"model.layers.11.attn.x_bias": "model-00002-of-00004.safetensors",
"model.layers.11.attn.x_proj.0.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.11.attn.x_proj.0.mu": "model-00002-of-00004.safetensors",
"model.layers.11.attn.x_proj.2.weight": "model-00002-of-00004.safetensors",
"model.layers.11.attn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.11.attn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.11.ffn.key.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.11.ffn.key.mu": "model-00002-of-00004.safetensors",
"model.layers.11.ffn.receptance.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.11.ffn.receptance.mu": "model-00002-of-00004.safetensors",
"model.layers.11.ffn.value.weight": "model-00002-of-00004.safetensors",
"model.layers.11.ffn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.11.ffn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.12.attn.bonus": "model-00002-of-00004.safetensors",
"model.layers.12.attn.g_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.12.attn.g_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.12.attn.g_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.12.attn.k_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.12.attn.o_proj.weight": "model-00002-of-00004.safetensors",
"model.layers.12.attn.r_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.12.attn.v_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.12.attn.w_proj.linear.lora.0.weight": "model-00002-of-00004.safetensors",
"model.layers.12.attn.w_proj.linear.lora.2.bias": "model-00002-of-00004.safetensors",
"model.layers.12.attn.w_proj.linear.lora.2.weight": "model-00002-of-00004.safetensors",
"model.layers.12.attn.x_bias": "model-00002-of-00004.safetensors",
"model.layers.12.attn.x_proj.0.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.12.attn.x_proj.0.mu": "model-00002-of-00004.safetensors",
"model.layers.12.attn.x_proj.2.weight": "model-00002-of-00004.safetensors",
"model.layers.12.attn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.12.attn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.12.ffn.key.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.12.ffn.key.mu": "model-00002-of-00004.safetensors",
"model.layers.12.ffn.receptance.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.12.ffn.receptance.mu": "model-00002-of-00004.safetensors",
"model.layers.12.ffn.value.weight": "model-00002-of-00004.safetensors",
"model.layers.12.ffn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.12.ffn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.13.attn.bonus": "model-00002-of-00004.safetensors",
"model.layers.13.attn.g_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.13.attn.g_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.13.attn.g_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.13.attn.k_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.13.attn.o_proj.weight": "model-00002-of-00004.safetensors",
"model.layers.13.attn.r_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.13.attn.v_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.13.attn.w_proj.linear.lora.0.weight": "model-00002-of-00004.safetensors",
"model.layers.13.attn.w_proj.linear.lora.2.bias": "model-00002-of-00004.safetensors",
"model.layers.13.attn.w_proj.linear.lora.2.weight": "model-00002-of-00004.safetensors",
"model.layers.13.attn.x_bias": "model-00002-of-00004.safetensors",
"model.layers.13.attn.x_proj.0.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.13.attn.x_proj.0.mu": "model-00002-of-00004.safetensors",
"model.layers.13.attn.x_proj.2.weight": "model-00002-of-00004.safetensors",
"model.layers.13.attn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.13.attn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.13.ffn.key.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.13.ffn.key.mu": "model-00002-of-00004.safetensors",
"model.layers.13.ffn.receptance.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.13.ffn.receptance.mu": "model-00002-of-00004.safetensors",
"model.layers.13.ffn.value.weight": "model-00002-of-00004.safetensors",
"model.layers.13.ffn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.13.ffn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.14.attn.bonus": "model-00002-of-00004.safetensors",
"model.layers.14.attn.g_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.14.attn.g_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.14.attn.g_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.14.attn.k_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.14.attn.o_proj.weight": "model-00002-of-00004.safetensors",
"model.layers.14.attn.r_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.14.attn.v_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.14.attn.w_proj.linear.lora.0.weight": "model-00002-of-00004.safetensors",
"model.layers.14.attn.w_proj.linear.lora.2.bias": "model-00002-of-00004.safetensors",
"model.layers.14.attn.w_proj.linear.lora.2.weight": "model-00002-of-00004.safetensors",
"model.layers.14.attn.x_bias": "model-00002-of-00004.safetensors",
"model.layers.14.attn.x_proj.0.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.14.attn.x_proj.0.mu": "model-00002-of-00004.safetensors",
"model.layers.14.attn.x_proj.2.weight": "model-00002-of-00004.safetensors",
"model.layers.14.attn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.14.attn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.14.ffn.key.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.14.ffn.key.mu": "model-00002-of-00004.safetensors",
"model.layers.14.ffn.receptance.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.14.ffn.receptance.mu": "model-00002-of-00004.safetensors",
"model.layers.14.ffn.value.weight": "model-00002-of-00004.safetensors",
"model.layers.14.ffn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.14.ffn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.15.attn.bonus": "model-00002-of-00004.safetensors",
"model.layers.15.attn.g_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.15.attn.g_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.15.attn.g_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.15.attn.k_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.15.attn.o_proj.weight": "model-00002-of-00004.safetensors",
"model.layers.15.attn.r_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.15.attn.v_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.15.attn.w_proj.linear.lora.0.weight": "model-00002-of-00004.safetensors",
"model.layers.15.attn.w_proj.linear.lora.2.bias": "model-00002-of-00004.safetensors",
"model.layers.15.attn.w_proj.linear.lora.2.weight": "model-00002-of-00004.safetensors",
"model.layers.15.attn.x_bias": "model-00002-of-00004.safetensors",
"model.layers.15.attn.x_proj.0.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.15.attn.x_proj.0.mu": "model-00002-of-00004.safetensors",
"model.layers.15.attn.x_proj.2.weight": "model-00002-of-00004.safetensors",
"model.layers.15.attn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.15.attn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.15.ffn.key.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.15.ffn.key.mu": "model-00002-of-00004.safetensors",
"model.layers.15.ffn.receptance.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.15.ffn.receptance.mu": "model-00002-of-00004.safetensors",
"model.layers.15.ffn.value.weight": "model-00002-of-00004.safetensors",
"model.layers.15.ffn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.15.ffn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.16.attn.bonus": "model-00002-of-00004.safetensors",
"model.layers.16.attn.g_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.16.attn.g_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.16.attn.g_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.16.attn.k_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.16.attn.o_proj.weight": "model-00002-of-00004.safetensors",
"model.layers.16.attn.r_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.16.attn.v_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.16.attn.w_proj.linear.lora.0.weight": "model-00002-of-00004.safetensors",
"model.layers.16.attn.w_proj.linear.lora.2.bias": "model-00002-of-00004.safetensors",
"model.layers.16.attn.w_proj.linear.lora.2.weight": "model-00002-of-00004.safetensors",
"model.layers.16.attn.x_bias": "model-00002-of-00004.safetensors",
"model.layers.16.attn.x_proj.0.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.16.attn.x_proj.0.mu": "model-00002-of-00004.safetensors",
"model.layers.16.attn.x_proj.2.weight": "model-00002-of-00004.safetensors",
"model.layers.16.attn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.16.attn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.16.ffn.key.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.16.ffn.key.mu": "model-00002-of-00004.safetensors",
"model.layers.16.ffn.receptance.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.16.ffn.receptance.mu": "model-00002-of-00004.safetensors",
"model.layers.16.ffn.value.weight": "model-00002-of-00004.safetensors",
"model.layers.16.ffn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.16.ffn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.17.attn.bonus": "model-00002-of-00004.safetensors",
"model.layers.17.attn.g_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.17.attn.g_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.17.attn.g_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.17.attn.k_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.17.attn.o_proj.weight": "model-00002-of-00004.safetensors",
"model.layers.17.attn.r_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.17.attn.v_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.17.attn.w_proj.linear.lora.0.weight": "model-00002-of-00004.safetensors",
"model.layers.17.attn.w_proj.linear.lora.2.bias": "model-00002-of-00004.safetensors",
"model.layers.17.attn.w_proj.linear.lora.2.weight": "model-00002-of-00004.safetensors",
"model.layers.17.attn.x_bias": "model-00002-of-00004.safetensors",
"model.layers.17.attn.x_proj.0.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.17.attn.x_proj.0.mu": "model-00002-of-00004.safetensors",
"model.layers.17.attn.x_proj.2.weight": "model-00002-of-00004.safetensors",
"model.layers.17.attn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.17.attn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.17.ffn.key.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.17.ffn.key.mu": "model-00002-of-00004.safetensors",
"model.layers.17.ffn.receptance.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.17.ffn.receptance.mu": "model-00002-of-00004.safetensors",
"model.layers.17.ffn.value.weight": "model-00002-of-00004.safetensors",
"model.layers.17.ffn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.17.ffn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.18.attn.bonus": "model-00002-of-00004.safetensors",
"model.layers.18.attn.g_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.18.attn.g_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.18.attn.g_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.18.attn.k_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.18.attn.o_proj.weight": "model-00002-of-00004.safetensors",
"model.layers.18.attn.r_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.18.attn.v_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.18.attn.w_proj.linear.lora.0.weight": "model-00002-of-00004.safetensors",
"model.layers.18.attn.w_proj.linear.lora.2.bias": "model-00002-of-00004.safetensors",
"model.layers.18.attn.w_proj.linear.lora.2.weight": "model-00002-of-00004.safetensors",
"model.layers.18.attn.x_bias": "model-00002-of-00004.safetensors",
"model.layers.18.attn.x_proj.0.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.18.attn.x_proj.0.mu": "model-00002-of-00004.safetensors",
"model.layers.18.attn.x_proj.2.weight": "model-00002-of-00004.safetensors",
"model.layers.18.attn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.18.attn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.18.ffn.key.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.18.ffn.key.mu": "model-00002-of-00004.safetensors",
"model.layers.18.ffn.receptance.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.18.ffn.receptance.mu": "model-00002-of-00004.safetensors",
"model.layers.18.ffn.value.weight": "model-00002-of-00004.safetensors",
"model.layers.18.ffn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.18.ffn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.19.attn.bonus": "model-00002-of-00004.safetensors",
"model.layers.19.attn.g_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.19.attn.g_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.19.attn.g_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.19.attn.k_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.19.attn.o_proj.weight": "model-00002-of-00004.safetensors",
"model.layers.19.attn.r_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.19.attn.v_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.19.attn.w_proj.linear.lora.0.weight": "model-00002-of-00004.safetensors",
"model.layers.19.attn.w_proj.linear.lora.2.bias": "model-00002-of-00004.safetensors",
"model.layers.19.attn.w_proj.linear.lora.2.weight": "model-00002-of-00004.safetensors",
"model.layers.19.attn.x_bias": "model-00002-of-00004.safetensors",
"model.layers.19.attn.x_proj.0.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.19.attn.x_proj.0.mu": "model-00002-of-00004.safetensors",
"model.layers.19.attn.x_proj.2.weight": "model-00002-of-00004.safetensors",
"model.layers.19.attn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.19.attn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.19.ffn.key.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.19.ffn.key.mu": "model-00002-of-00004.safetensors",
"model.layers.19.ffn.receptance.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.19.ffn.receptance.mu": "model-00002-of-00004.safetensors",
"model.layers.19.ffn.value.weight": "model-00002-of-00004.safetensors",
"model.layers.19.ffn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.19.ffn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.2.attn.bonus": "model-00001-of-00004.safetensors",
"model.layers.2.attn.g_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.2.attn.g_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.2.attn.g_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.2.attn.k_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.2.attn.o_proj.weight": "model-00001-of-00004.safetensors",
"model.layers.2.attn.r_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.2.attn.v_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.2.attn.w_proj.linear.lora.0.weight": "model-00001-of-00004.safetensors",
"model.layers.2.attn.w_proj.linear.lora.2.bias": "model-00001-of-00004.safetensors",
"model.layers.2.attn.w_proj.linear.lora.2.weight": "model-00001-of-00004.safetensors",
"model.layers.2.attn.x_bias": "model-00001-of-00004.safetensors",
"model.layers.2.attn.x_proj.0.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.2.attn.x_proj.0.mu": "model-00001-of-00004.safetensors",
"model.layers.2.attn.x_proj.2.weight": "model-00001-of-00004.safetensors",
"model.layers.2.attn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.2.attn_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.2.ffn.key.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.2.ffn.key.mu": "model-00001-of-00004.safetensors",
"model.layers.2.ffn.receptance.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.2.ffn.receptance.mu": "model-00001-of-00004.safetensors",
"model.layers.2.ffn.value.weight": "model-00001-of-00004.safetensors",
"model.layers.2.ffn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.2.ffn_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.20.attn.bonus": "model-00002-of-00004.safetensors",
"model.layers.20.attn.g_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.20.attn.g_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.20.attn.g_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.20.attn.k_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.20.attn.o_proj.weight": "model-00002-of-00004.safetensors",
"model.layers.20.attn.r_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.20.attn.v_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.20.attn.w_proj.linear.lora.0.weight": "model-00002-of-00004.safetensors",
"model.layers.20.attn.w_proj.linear.lora.2.bias": "model-00002-of-00004.safetensors",
"model.layers.20.attn.w_proj.linear.lora.2.weight": "model-00002-of-00004.safetensors",
"model.layers.20.attn.x_bias": "model-00002-of-00004.safetensors",
"model.layers.20.attn.x_proj.0.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.20.attn.x_proj.0.mu": "model-00002-of-00004.safetensors",
"model.layers.20.attn.x_proj.2.weight": "model-00002-of-00004.safetensors",
"model.layers.20.attn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.20.attn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.20.ffn.key.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.20.ffn.key.mu": "model-00002-of-00004.safetensors",
"model.layers.20.ffn.receptance.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.20.ffn.receptance.mu": "model-00002-of-00004.safetensors",
"model.layers.20.ffn.value.weight": "model-00002-of-00004.safetensors",
"model.layers.20.ffn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.20.ffn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.21.attn.bonus": "model-00002-of-00004.safetensors",
"model.layers.21.attn.g_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.21.attn.g_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.21.attn.g_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.21.attn.k_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.21.attn.o_proj.weight": "model-00003-of-00004.safetensors",
"model.layers.21.attn.r_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.21.attn.v_proj.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.21.attn.w_proj.linear.lora.0.weight": "model-00002-of-00004.safetensors",
"model.layers.21.attn.w_proj.linear.lora.2.bias": "model-00002-of-00004.safetensors",
"model.layers.21.attn.w_proj.linear.lora.2.weight": "model-00002-of-00004.safetensors",
"model.layers.21.attn.x_bias": "model-00002-of-00004.safetensors",
"model.layers.21.attn.x_proj.0.linear.weight": "model-00002-of-00004.safetensors",
"model.layers.21.attn.x_proj.0.mu": "model-00002-of-00004.safetensors",
"model.layers.21.attn.x_proj.2.weight": "model-00002-of-00004.safetensors",
"model.layers.21.attn_norm.bias": "model-00002-of-00004.safetensors",
"model.layers.21.attn_norm.weight": "model-00002-of-00004.safetensors",
"model.layers.21.ffn.key.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.21.ffn.key.mu": "model-00003-of-00004.safetensors",
"model.layers.21.ffn.receptance.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.21.ffn.receptance.mu": "model-00003-of-00004.safetensors",
"model.layers.21.ffn.value.weight": "model-00003-of-00004.safetensors",
"model.layers.21.ffn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.21.ffn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.22.attn.bonus": "model-00003-of-00004.safetensors",
"model.layers.22.attn.g_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.22.attn.g_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.22.attn.g_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.22.attn.k_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.22.attn.o_proj.weight": "model-00003-of-00004.safetensors",
"model.layers.22.attn.r_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.22.attn.v_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.22.attn.w_proj.linear.lora.0.weight": "model-00003-of-00004.safetensors",
"model.layers.22.attn.w_proj.linear.lora.2.bias": "model-00003-of-00004.safetensors",
"model.layers.22.attn.w_proj.linear.lora.2.weight": "model-00003-of-00004.safetensors",
"model.layers.22.attn.x_bias": "model-00003-of-00004.safetensors",
"model.layers.22.attn.x_proj.0.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.22.attn.x_proj.0.mu": "model-00003-of-00004.safetensors",
"model.layers.22.attn.x_proj.2.weight": "model-00003-of-00004.safetensors",
"model.layers.22.attn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.22.attn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.22.ffn.key.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.22.ffn.key.mu": "model-00003-of-00004.safetensors",
"model.layers.22.ffn.receptance.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.22.ffn.receptance.mu": "model-00003-of-00004.safetensors",
"model.layers.22.ffn.value.weight": "model-00003-of-00004.safetensors",
"model.layers.22.ffn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.22.ffn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.23.attn.bonus": "model-00003-of-00004.safetensors",
"model.layers.23.attn.g_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.23.attn.g_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.23.attn.g_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.23.attn.k_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.23.attn.o_proj.weight": "model-00003-of-00004.safetensors",
"model.layers.23.attn.r_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.23.attn.v_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.23.attn.w_proj.linear.lora.0.weight": "model-00003-of-00004.safetensors",
"model.layers.23.attn.w_proj.linear.lora.2.bias": "model-00003-of-00004.safetensors",
"model.layers.23.attn.w_proj.linear.lora.2.weight": "model-00003-of-00004.safetensors",
"model.layers.23.attn.x_bias": "model-00003-of-00004.safetensors",
"model.layers.23.attn.x_proj.0.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.23.attn.x_proj.0.mu": "model-00003-of-00004.safetensors",
"model.layers.23.attn.x_proj.2.weight": "model-00003-of-00004.safetensors",
"model.layers.23.attn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.23.attn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.23.ffn.key.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.23.ffn.key.mu": "model-00003-of-00004.safetensors",
"model.layers.23.ffn.receptance.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.23.ffn.receptance.mu": "model-00003-of-00004.safetensors",
"model.layers.23.ffn.value.weight": "model-00003-of-00004.safetensors",
"model.layers.23.ffn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.23.ffn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.24.attn.bonus": "model-00003-of-00004.safetensors",
"model.layers.24.attn.g_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.24.attn.g_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.24.attn.g_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.24.attn.k_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.24.attn.o_proj.weight": "model-00003-of-00004.safetensors",
"model.layers.24.attn.r_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.24.attn.v_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.24.attn.w_proj.linear.lora.0.weight": "model-00003-of-00004.safetensors",
"model.layers.24.attn.w_proj.linear.lora.2.bias": "model-00003-of-00004.safetensors",
"model.layers.24.attn.w_proj.linear.lora.2.weight": "model-00003-of-00004.safetensors",
"model.layers.24.attn.x_bias": "model-00003-of-00004.safetensors",
"model.layers.24.attn.x_proj.0.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.24.attn.x_proj.0.mu": "model-00003-of-00004.safetensors",
"model.layers.24.attn.x_proj.2.weight": "model-00003-of-00004.safetensors",
"model.layers.24.attn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.24.attn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.24.ffn.key.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.24.ffn.key.mu": "model-00003-of-00004.safetensors",
"model.layers.24.ffn.receptance.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.24.ffn.receptance.mu": "model-00003-of-00004.safetensors",
"model.layers.24.ffn.value.weight": "model-00003-of-00004.safetensors",
"model.layers.24.ffn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.24.ffn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.25.attn.bonus": "model-00003-of-00004.safetensors",
"model.layers.25.attn.g_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.25.attn.g_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.25.attn.g_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.25.attn.k_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.25.attn.o_proj.weight": "model-00003-of-00004.safetensors",
"model.layers.25.attn.r_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.25.attn.v_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.25.attn.w_proj.linear.lora.0.weight": "model-00003-of-00004.safetensors",
"model.layers.25.attn.w_proj.linear.lora.2.bias": "model-00003-of-00004.safetensors",
"model.layers.25.attn.w_proj.linear.lora.2.weight": "model-00003-of-00004.safetensors",
"model.layers.25.attn.x_bias": "model-00003-of-00004.safetensors",
"model.layers.25.attn.x_proj.0.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.25.attn.x_proj.0.mu": "model-00003-of-00004.safetensors",
"model.layers.25.attn.x_proj.2.weight": "model-00003-of-00004.safetensors",
"model.layers.25.attn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.25.attn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.25.ffn.key.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.25.ffn.key.mu": "model-00003-of-00004.safetensors",
"model.layers.25.ffn.receptance.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.25.ffn.receptance.mu": "model-00003-of-00004.safetensors",
"model.layers.25.ffn.value.weight": "model-00003-of-00004.safetensors",
"model.layers.25.ffn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.25.ffn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.26.attn.bonus": "model-00003-of-00004.safetensors",
"model.layers.26.attn.g_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.26.attn.g_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.26.attn.g_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.26.attn.k_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.26.attn.o_proj.weight": "model-00003-of-00004.safetensors",
"model.layers.26.attn.r_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.26.attn.v_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.26.attn.w_proj.linear.lora.0.weight": "model-00003-of-00004.safetensors",
"model.layers.26.attn.w_proj.linear.lora.2.bias": "model-00003-of-00004.safetensors",
"model.layers.26.attn.w_proj.linear.lora.2.weight": "model-00003-of-00004.safetensors",
"model.layers.26.attn.x_bias": "model-00003-of-00004.safetensors",
"model.layers.26.attn.x_proj.0.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.26.attn.x_proj.0.mu": "model-00003-of-00004.safetensors",
"model.layers.26.attn.x_proj.2.weight": "model-00003-of-00004.safetensors",
"model.layers.26.attn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.26.attn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.26.ffn.key.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.26.ffn.key.mu": "model-00003-of-00004.safetensors",
"model.layers.26.ffn.receptance.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.26.ffn.receptance.mu": "model-00003-of-00004.safetensors",
"model.layers.26.ffn.value.weight": "model-00003-of-00004.safetensors",
"model.layers.26.ffn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.26.ffn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.27.attn.bonus": "model-00003-of-00004.safetensors",
"model.layers.27.attn.g_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.27.attn.g_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.27.attn.g_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.27.attn.k_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.27.attn.o_proj.weight": "model-00003-of-00004.safetensors",
"model.layers.27.attn.r_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.27.attn.v_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.27.attn.w_proj.linear.lora.0.weight": "model-00003-of-00004.safetensors",
"model.layers.27.attn.w_proj.linear.lora.2.bias": "model-00003-of-00004.safetensors",
"model.layers.27.attn.w_proj.linear.lora.2.weight": "model-00003-of-00004.safetensors",
"model.layers.27.attn.x_bias": "model-00003-of-00004.safetensors",
"model.layers.27.attn.x_proj.0.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.27.attn.x_proj.0.mu": "model-00003-of-00004.safetensors",
"model.layers.27.attn.x_proj.2.weight": "model-00003-of-00004.safetensors",
"model.layers.27.attn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.27.attn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.27.ffn.key.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.27.ffn.key.mu": "model-00003-of-00004.safetensors",
"model.layers.27.ffn.receptance.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.27.ffn.receptance.mu": "model-00003-of-00004.safetensors",
"model.layers.27.ffn.value.weight": "model-00003-of-00004.safetensors",
"model.layers.27.ffn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.27.ffn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.28.attn.bonus": "model-00003-of-00004.safetensors",
"model.layers.28.attn.g_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.28.attn.g_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.28.attn.g_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.28.attn.k_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.28.attn.o_proj.weight": "model-00003-of-00004.safetensors",
"model.layers.28.attn.r_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.28.attn.v_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.28.attn.w_proj.linear.lora.0.weight": "model-00003-of-00004.safetensors",
"model.layers.28.attn.w_proj.linear.lora.2.bias": "model-00003-of-00004.safetensors",
"model.layers.28.attn.w_proj.linear.lora.2.weight": "model-00003-of-00004.safetensors",
"model.layers.28.attn.x_bias": "model-00003-of-00004.safetensors",
"model.layers.28.attn.x_proj.0.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.28.attn.x_proj.0.mu": "model-00003-of-00004.safetensors",
"model.layers.28.attn.x_proj.2.weight": "model-00003-of-00004.safetensors",
"model.layers.28.attn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.28.attn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.28.ffn.key.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.28.ffn.key.mu": "model-00003-of-00004.safetensors",
"model.layers.28.ffn.receptance.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.28.ffn.receptance.mu": "model-00003-of-00004.safetensors",
"model.layers.28.ffn.value.weight": "model-00003-of-00004.safetensors",
"model.layers.28.ffn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.28.ffn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.29.attn.bonus": "model-00003-of-00004.safetensors",
"model.layers.29.attn.g_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.29.attn.g_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.29.attn.g_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.29.attn.k_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.29.attn.o_proj.weight": "model-00003-of-00004.safetensors",
"model.layers.29.attn.r_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.29.attn.v_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.29.attn.w_proj.linear.lora.0.weight": "model-00003-of-00004.safetensors",
"model.layers.29.attn.w_proj.linear.lora.2.bias": "model-00003-of-00004.safetensors",
"model.layers.29.attn.w_proj.linear.lora.2.weight": "model-00003-of-00004.safetensors",
"model.layers.29.attn.x_bias": "model-00003-of-00004.safetensors",
"model.layers.29.attn.x_proj.0.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.29.attn.x_proj.0.mu": "model-00003-of-00004.safetensors",
"model.layers.29.attn.x_proj.2.weight": "model-00003-of-00004.safetensors",
"model.layers.29.attn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.29.attn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.29.ffn.key.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.29.ffn.key.mu": "model-00003-of-00004.safetensors",
"model.layers.29.ffn.receptance.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.29.ffn.receptance.mu": "model-00003-of-00004.safetensors",
"model.layers.29.ffn.value.weight": "model-00003-of-00004.safetensors",
"model.layers.29.ffn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.29.ffn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.3.attn.bonus": "model-00001-of-00004.safetensors",
"model.layers.3.attn.g_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.3.attn.g_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.3.attn.g_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.3.attn.k_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.3.attn.o_proj.weight": "model-00001-of-00004.safetensors",
"model.layers.3.attn.r_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.3.attn.v_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.3.attn.w_proj.linear.lora.0.weight": "model-00001-of-00004.safetensors",
"model.layers.3.attn.w_proj.linear.lora.2.bias": "model-00001-of-00004.safetensors",
"model.layers.3.attn.w_proj.linear.lora.2.weight": "model-00001-of-00004.safetensors",
"model.layers.3.attn.x_bias": "model-00001-of-00004.safetensors",
"model.layers.3.attn.x_proj.0.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.3.attn.x_proj.0.mu": "model-00001-of-00004.safetensors",
"model.layers.3.attn.x_proj.2.weight": "model-00001-of-00004.safetensors",
"model.layers.3.attn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.3.attn_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.3.ffn.key.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.3.ffn.key.mu": "model-00001-of-00004.safetensors",
"model.layers.3.ffn.receptance.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.3.ffn.receptance.mu": "model-00001-of-00004.safetensors",
"model.layers.3.ffn.value.weight": "model-00001-of-00004.safetensors",
"model.layers.3.ffn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.3.ffn_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.30.attn.bonus": "model-00003-of-00004.safetensors",
"model.layers.30.attn.g_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.30.attn.g_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.30.attn.g_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.30.attn.k_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.30.attn.o_proj.weight": "model-00003-of-00004.safetensors",
"model.layers.30.attn.r_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.30.attn.v_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.30.attn.w_proj.linear.lora.0.weight": "model-00003-of-00004.safetensors",
"model.layers.30.attn.w_proj.linear.lora.2.bias": "model-00003-of-00004.safetensors",
"model.layers.30.attn.w_proj.linear.lora.2.weight": "model-00003-of-00004.safetensors",
"model.layers.30.attn.x_bias": "model-00003-of-00004.safetensors",
"model.layers.30.attn.x_proj.0.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.30.attn.x_proj.0.mu": "model-00003-of-00004.safetensors",
"model.layers.30.attn.x_proj.2.weight": "model-00003-of-00004.safetensors",
"model.layers.30.attn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.30.attn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.30.ffn.key.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.30.ffn.key.mu": "model-00003-of-00004.safetensors",
"model.layers.30.ffn.receptance.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.30.ffn.receptance.mu": "model-00003-of-00004.safetensors",
"model.layers.30.ffn.value.weight": "model-00003-of-00004.safetensors",
"model.layers.30.ffn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.30.ffn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.31.attn.bonus": "model-00003-of-00004.safetensors",
"model.layers.31.attn.g_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.31.attn.g_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.31.attn.g_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.31.attn.k_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.31.attn.o_proj.weight": "model-00003-of-00004.safetensors",
"model.layers.31.attn.r_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.31.attn.v_proj.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.31.attn.w_proj.linear.lora.0.weight": "model-00003-of-00004.safetensors",
"model.layers.31.attn.w_proj.linear.lora.2.bias": "model-00003-of-00004.safetensors",
"model.layers.31.attn.w_proj.linear.lora.2.weight": "model-00003-of-00004.safetensors",
"model.layers.31.attn.x_bias": "model-00003-of-00004.safetensors",
"model.layers.31.attn.x_proj.0.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.31.attn.x_proj.0.mu": "model-00003-of-00004.safetensors",
"model.layers.31.attn.x_proj.2.weight": "model-00003-of-00004.safetensors",
"model.layers.31.attn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.31.attn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.31.ffn.key.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.31.ffn.key.mu": "model-00003-of-00004.safetensors",
"model.layers.31.ffn.receptance.linear.weight": "model-00003-of-00004.safetensors",
"model.layers.31.ffn.receptance.mu": "model-00003-of-00004.safetensors",
"model.layers.31.ffn.value.weight": "model-00003-of-00004.safetensors",
"model.layers.31.ffn_norm.bias": "model-00003-of-00004.safetensors",
"model.layers.31.ffn_norm.weight": "model-00003-of-00004.safetensors",
"model.layers.4.attn.bonus": "model-00001-of-00004.safetensors",
"model.layers.4.attn.g_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.4.attn.g_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.4.attn.g_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.4.attn.k_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.4.attn.o_proj.weight": "model-00001-of-00004.safetensors",
"model.layers.4.attn.r_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.4.attn.v_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.4.attn.w_proj.linear.lora.0.weight": "model-00001-of-00004.safetensors",
"model.layers.4.attn.w_proj.linear.lora.2.bias": "model-00001-of-00004.safetensors",
"model.layers.4.attn.w_proj.linear.lora.2.weight": "model-00001-of-00004.safetensors",
"model.layers.4.attn.x_bias": "model-00001-of-00004.safetensors",
"model.layers.4.attn.x_proj.0.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.4.attn.x_proj.0.mu": "model-00001-of-00004.safetensors",
"model.layers.4.attn.x_proj.2.weight": "model-00001-of-00004.safetensors",
"model.layers.4.attn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.4.attn_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.4.ffn.key.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.4.ffn.key.mu": "model-00001-of-00004.safetensors",
"model.layers.4.ffn.receptance.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.4.ffn.receptance.mu": "model-00001-of-00004.safetensors",
"model.layers.4.ffn.value.weight": "model-00001-of-00004.safetensors",
"model.layers.4.ffn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.4.ffn_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.5.attn.bonus": "model-00001-of-00004.safetensors",
"model.layers.5.attn.g_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.5.attn.g_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.5.attn.g_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.5.attn.k_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.5.attn.o_proj.weight": "model-00001-of-00004.safetensors",
"model.layers.5.attn.r_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.5.attn.v_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.5.attn.w_proj.linear.lora.0.weight": "model-00001-of-00004.safetensors",
"model.layers.5.attn.w_proj.linear.lora.2.bias": "model-00001-of-00004.safetensors",
"model.layers.5.attn.w_proj.linear.lora.2.weight": "model-00001-of-00004.safetensors",
"model.layers.5.attn.x_bias": "model-00001-of-00004.safetensors",
"model.layers.5.attn.x_proj.0.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.5.attn.x_proj.0.mu": "model-00001-of-00004.safetensors",
"model.layers.5.attn.x_proj.2.weight": "model-00001-of-00004.safetensors",
"model.layers.5.attn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.5.attn_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.5.ffn.key.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.5.ffn.key.mu": "model-00001-of-00004.safetensors",
"model.layers.5.ffn.receptance.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.5.ffn.receptance.mu": "model-00001-of-00004.safetensors",
"model.layers.5.ffn.value.weight": "model-00001-of-00004.safetensors",
"model.layers.5.ffn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.5.ffn_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.6.attn.bonus": "model-00001-of-00004.safetensors",
"model.layers.6.attn.g_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.6.attn.g_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.6.attn.g_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.6.attn.k_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.6.attn.o_proj.weight": "model-00001-of-00004.safetensors",
"model.layers.6.attn.r_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.6.attn.v_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.6.attn.w_proj.linear.lora.0.weight": "model-00001-of-00004.safetensors",
"model.layers.6.attn.w_proj.linear.lora.2.bias": "model-00001-of-00004.safetensors",
"model.layers.6.attn.w_proj.linear.lora.2.weight": "model-00001-of-00004.safetensors",
"model.layers.6.attn.x_bias": "model-00001-of-00004.safetensors",
"model.layers.6.attn.x_proj.0.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.6.attn.x_proj.0.mu": "model-00001-of-00004.safetensors",
"model.layers.6.attn.x_proj.2.weight": "model-00001-of-00004.safetensors",
"model.layers.6.attn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.6.attn_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.6.ffn.key.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.6.ffn.key.mu": "model-00001-of-00004.safetensors",
"model.layers.6.ffn.receptance.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.6.ffn.receptance.mu": "model-00001-of-00004.safetensors",
"model.layers.6.ffn.value.weight": "model-00001-of-00004.safetensors",
"model.layers.6.ffn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.6.ffn_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.7.attn.bonus": "model-00001-of-00004.safetensors",
"model.layers.7.attn.g_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.7.attn.g_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.7.attn.g_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.7.attn.k_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.7.attn.o_proj.weight": "model-00001-of-00004.safetensors",
"model.layers.7.attn.r_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.7.attn.v_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.7.attn.w_proj.linear.lora.0.weight": "model-00001-of-00004.safetensors",
"model.layers.7.attn.w_proj.linear.lora.2.bias": "model-00001-of-00004.safetensors",
"model.layers.7.attn.w_proj.linear.lora.2.weight": "model-00001-of-00004.safetensors",
"model.layers.7.attn.x_bias": "model-00001-of-00004.safetensors",
"model.layers.7.attn.x_proj.0.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.7.attn.x_proj.0.mu": "model-00001-of-00004.safetensors",
"model.layers.7.attn.x_proj.2.weight": "model-00001-of-00004.safetensors",
"model.layers.7.attn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.7.attn_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.7.ffn.key.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.7.ffn.key.mu": "model-00001-of-00004.safetensors",
"model.layers.7.ffn.receptance.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.7.ffn.receptance.mu": "model-00001-of-00004.safetensors",
"model.layers.7.ffn.value.weight": "model-00001-of-00004.safetensors",
"model.layers.7.ffn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.7.ffn_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.8.attn.bonus": "model-00001-of-00004.safetensors",
"model.layers.8.attn.g_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.8.attn.g_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.8.attn.g_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.8.attn.k_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.8.attn.o_proj.weight": "model-00001-of-00004.safetensors",
"model.layers.8.attn.r_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.8.attn.v_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.8.attn.w_proj.linear.lora.0.weight": "model-00001-of-00004.safetensors",
"model.layers.8.attn.w_proj.linear.lora.2.bias": "model-00001-of-00004.safetensors",
"model.layers.8.attn.w_proj.linear.lora.2.weight": "model-00001-of-00004.safetensors",
"model.layers.8.attn.x_bias": "model-00001-of-00004.safetensors",
"model.layers.8.attn.x_proj.0.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.8.attn.x_proj.0.mu": "model-00001-of-00004.safetensors",
"model.layers.8.attn.x_proj.2.weight": "model-00001-of-00004.safetensors",
"model.layers.8.attn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.8.attn_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.8.ffn.key.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.8.ffn.key.mu": "model-00001-of-00004.safetensors",
"model.layers.8.ffn.receptance.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.8.ffn.receptance.mu": "model-00001-of-00004.safetensors",
"model.layers.8.ffn.value.weight": "model-00001-of-00004.safetensors",
"model.layers.8.ffn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.8.ffn_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.9.attn.bonus": "model-00001-of-00004.safetensors",
"model.layers.9.attn.g_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.9.attn.g_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.9.attn.g_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.9.attn.k_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.9.attn.o_proj.weight": "model-00001-of-00004.safetensors",
"model.layers.9.attn.r_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.9.attn.v_proj.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.9.attn.w_proj.linear.lora.0.weight": "model-00001-of-00004.safetensors",
"model.layers.9.attn.w_proj.linear.lora.2.bias": "model-00001-of-00004.safetensors",
"model.layers.9.attn.w_proj.linear.lora.2.weight": "model-00001-of-00004.safetensors",
"model.layers.9.attn.x_bias": "model-00001-of-00004.safetensors",
"model.layers.9.attn.x_proj.0.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.9.attn.x_proj.0.mu": "model-00001-of-00004.safetensors",
"model.layers.9.attn.x_proj.2.weight": "model-00001-of-00004.safetensors",
"model.layers.9.attn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.9.attn_norm.weight": "model-00001-of-00004.safetensors",
"model.layers.9.ffn.key.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.9.ffn.key.mu": "model-00001-of-00004.safetensors",
"model.layers.9.ffn.receptance.linear.weight": "model-00001-of-00004.safetensors",
"model.layers.9.ffn.receptance.mu": "model-00001-of-00004.safetensors",
"model.layers.9.ffn.value.weight": "model-00001-of-00004.safetensors",
"model.layers.9.ffn_norm.bias": "model-00001-of-00004.safetensors",
"model.layers.9.ffn_norm.weight": "model-00001-of-00004.safetensors",
"model.norm.bias": "model-00003-of-00004.safetensors",
"model.norm.weight": "model-00003-of-00004.safetensors"
}
}