charlesdedampierre commited on
Commit
c1c3fdd
1 Parent(s): 2b0a714

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -2
README.md CHANGED
@@ -5,9 +5,8 @@ license: apache-2.0
5
  ## Model description
6
 
7
 
8
- TopicNeuralHermes 2.5 Mistral 7B is a refined model developed through fine-tuning with a specific subset of data, selected via Topic Modeling Techniques using [Bunkatopics](https://github.com/charlesdedampierre/BunkaTopics).
9
 
10
- continuing from OpenHermes 2.5.
11
 
12
  The model was trained on a refined DPO dataset. The objective was to train the model on a small portion of the DPO data. To achieve this, we compared two datasets used to train the reward model: the rejected Llama answers and the accepted ChatGPT answers from the [DPO dataset](mlabonne/chatml_dpo_pairs).
13
  We then conducted topic modeling on both datasets, keeping only the topics that existed in the accepted dataset but not in the rejected one.
 
5
  ## Model description
6
 
7
 
8
+ TopicNeuralHermes 2.5 Mistral 7B is a refined model developed through fine-tuning with a specific subset of data, selected via Topic Modeling Techniques using [Bunkatopics](https://github.com/charlesdedampierre/BunkaTopics), as a continuing from [OpenHermes 2.5](teknium/OpenHermes-2.5-Mistral-7B).
9
 
 
10
 
11
  The model was trained on a refined DPO dataset. The objective was to train the model on a small portion of the DPO data. To achieve this, we compared two datasets used to train the reward model: the rejected Llama answers and the accepted ChatGPT answers from the [DPO dataset](mlabonne/chatml_dpo_pairs).
12
  We then conducted topic modeling on both datasets, keeping only the topics that existed in the accepted dataset but not in the rejected one.