|
--- |
|
license: apache-2.0 |
|
--- |
|
|
|
# etri-xainlp/SOLAR-10.7B-merge-dpo |
|
|
|
## Model Details |
|
|
|
**Model Developers** ETRI xainlp team |
|
|
|
**Input** text only. |
|
|
|
**Output** text only. |
|
|
|
**Model Architecture** |
|
|
|
We used MergeKit to merge Model heavytail/kullm-solar into Model upstage/SOLAR-10.7B-Instruct-v1.0 as the base. |
|
|
|
**Base Model** [upstage/SOLAR-10.7B-Instruct-v1.0](https://huggingface.co/upstage/SOLAR-10.7B-Instruct-v1.0) |
|
|
|
**Merge Model** [heavytail/kullm-solar](https://huggingface.co/heavytail/kullm-solar) |
|
|
|
**Training Dataset** |
|
|
|
- dpo+lora: 90k user preference set |
|
|
|
- We use A100 GPU 80GB * 1, when training. |