{ "config": { "adapter_residual_before_ln": false, "cross_adapter": false, "inv_adapter": null, "inv_adapter_reduction_factor": null, "leave_out": [], "ln_after": false, "ln_before": false, "mh_adapter": false, "non_linearity": "relu", "original_ln_after": true, "original_ln_before": true, "output_adapter": true, "reduction_factor": 16, "residual_before_ln": true }, "hidden_size": 768, "model_class": "RobertaModelWithHeads", "model_name": "roberta-base", "model_type": "roberta", "name": "glue_qqp" }