| { | |
| "archi_type": "decoder", | |
| "attn_implementation": "flash_attention_2", | |
| "attn_type": "causal", | |
| "bidirectional": true, | |
| "lexical": false, | |
| "lora": true, | |
| "lora_r": 64, | |
| "model_name_or_path": "Qwen/Qwen3-8B", | |
| "model_type": "splade", | |
| "n_layers": null, | |
| "padding_side": "left", | |
| "train_head": false, | |
| "transformers_version": "4.53.3", | |
| "auto_map": { | |
| "AutoConfig": "splade.SpladeConfig", | |
| "AutoModelForCausalLM": "splade.Splade" | |
| } | |
| } |