clee84 commited on
Commit
0875e77
1 Parent(s): 5830962

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -5
README.md CHANGED
@@ -15,12 +15,19 @@ tags:
15
  - NLP
16
  - DPO
17
  ---
18
- # Model Card for Mistral-7B-lighthouse-merge-v0.1
19
- This model is a result of merging two models A and B.
20
- The method used for merging is "slerp" with [mergekit](https://github.com/cg123/mergekit).
21
 
22
- A: [mistralai/Mistral-7B-instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2)
23
- B: [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) fine-tuned through DPO training.
 
 
 
 
 
 
 
 
 
24
 
25
  ### Jinja Prompt Template
26
  ```
 
15
  - NLP
16
  - DPO
17
  ---
18
+ # Model Overview
 
 
19
 
20
+ This model is a result of a sophisticated merging process involving two distinct models, Model A and Model B. The merging methodology employed is the "slerp" technique, facilitated by the use of [mergekit](https://github.com/cg123/mergekit).
21
+
22
+ ## Component Models
23
+
24
+ ### Model A
25
+ - **Source**: [mistralai/Mistral-7B-instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2)
26
+ - **Description**: This model is designed to provide instruction-based outputs, enhancing the clarity and precision in response generation.
27
+
28
+ ### Model B
29
+ - **Source**: Based on [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1)
30
+ - **Enhancements**: Fine-tuned using DPO (Direct Preference Optimization) training, augmenting its capabilities for more adaptive and context-aware responses.
31
 
32
  ### Jinja Prompt Template
33
  ```