{ "_attn_implementation_autoset": true, "attn_implementation": null, "auto_map": { "AutoConfig": "modelling_pisco.COCOMConfig", "AutoModel": "modelling_pisco.COCOM" }, "compr_base_model_name": null, "compr_bidirectional": false, "compr_every_n_layer": null, "compr_mlp_hidden_dim": 8096, "compr_model_name": "meta-llama/Llama-3.2-1B-Instruct", "compr_n_layers": null, "compr_rate": 16, "compr_rms_norm": false, "compr_use_mlp": true, "decoder_model_name": "mistralai/Mistral-7B-Instruct-v0.2", "device_map": "auto", "different_mem_tokens": true, "doc_max_length": 128, "generation_top_k": 1, "kbtc_training": false, "load_adapters": true, "lora": true, "lora_compressor": false, "lora_r": 16, "lora_r_compressor": 16, "model_type": "COCOM", "optimize_mem_tokens": true, "quantization": "no", "sep": true, "training_form": "both_separately", "transformers_version": "4.48.0" }