--- base_model: [intervitens/internlm2-base-20b-llama] language: - en pipeline_tag: text-generation tags: - mergekit - merge license: other datasets: - lemonilia/LimaRP --- Experimental model, LimaRP LoRA trained on top of internlm2-base-20b with 8192 context length and merged with internlm2-chat-20b. Prompt format is ChatML. *** # internlm2-06limarp-1chat-TASK_ARITHM-20b-v0.03 This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using intervitens/internlm2-base-20b-llama as a base. ### Models Merged The following models were included in the merge: * ./internlm2-limarp-20b-v0.03 * ./internlm2-chat-20b-llama ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: ./internlm2-chat-20b-llama parameters: weight: 1.0 - model: ./internlm2-limarp-20b-v0.03 parameters: weight: 0.6 merge_method: task_arithmetic base_model: ./internlm2-base-20b-llama parameters: #normalize: false #int8_mask: true dtype: bfloat16 ```