mlabonne commited on
Commit
33f76dd
β€’
1 Parent(s): 68ff49c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -2
README.md CHANGED
@@ -10,9 +10,11 @@ tags:
10
  - rlhf
11
  ---
12
 
 
 
13
  # NeuralBeagle14-7B
14
 
15
- NeuralBeagle14-7B is a DPO fine-tune of [mlabonne/Beagle14-7B](https://huggingface.co/mlabonne/Beagle14-7B) using [argilla/distilabel-intel-orca-dpo-pairs](https://huggingface.co/datasets/argilla/distilabel-intel-orca-dpo-pairs).
16
 
17
  Thanks [Argilla](https://huggingface.co/argilla) for providing the dataset and the training recipe [here](https://huggingface.co/argilla/distilabeled-Marcoro14-7B-slerp). πŸ’ͺ
18
 
@@ -22,7 +24,7 @@ The evaluation was performed using [LLM AutoEval](https://github.com/mlabonne/ll
22
 
23
  | Model | Average | AGIEval | GPT4All | TruthfulQA | Bigbench |
24
  |---|---:|---:|---:|---:|---:|
25
- |[**Beagle14-7B**](https://huggingface.co/mlabonne/Beagle14-7B)| ****| ****| ****| ****| **4**|
26
  | [mlabonne/Beagle14-7B](https://huggingface.co/mlabonne/Beagle14-7B) [πŸ“„](https://gist.github.com/mlabonne/f5a5bf8c0827bbec2f05b97cc62d642c) | 59.4 | 44.38 | 76.53 | 69.44 | 47.25 |
27
  | [mlabonne/NeuralDaredevil-7B](https://huggingface.co/mlabonne/NeuralDaredevil-7B) [πŸ“„](https://gist.github.com/mlabonne/cbeb077d1df71cb81c78f742f19f4155) | 59.39 | 45.23 | 76.2 | 67.61 | 48.52 |
28
  | [argilla/distilabeled-Marcoro14-7B-slerp](https://huggingface.co/argilla/distilabeled-Marcoro14-7B-slerp) [πŸ“„](https://gist.github.com/mlabonne/9082c4e59f4d3f3543c5eda3f4807040) | 58.93 | 45.38 | 76.48 | 65.68 | 48.18 |
 
10
  - rlhf
11
  ---
12
 
13
+ ![](https://i.imgur.com/89ZAKcn.png)
14
+
15
  # NeuralBeagle14-7B
16
 
17
+ NeuralBeagle14-7B is a DPO fine-tune of [mlabonne/Beagle14-7B](https://huggingface.co/mlabonne/Beagle14-7B) using the [argilla/distilabel-intel-orca-dpo-pairs](https://huggingface.co/datasets/argilla/distilabel-intel-orca-dpo-pairs) preference dataset and my DPO notebook from [this article](https://towardsdatascience.com/fine-tune-a-mistral-7b-model-with-direct-preference-optimization-708042745aac).
18
 
19
  Thanks [Argilla](https://huggingface.co/argilla) for providing the dataset and the training recipe [here](https://huggingface.co/argilla/distilabeled-Marcoro14-7B-slerp). πŸ’ͺ
20
 
 
24
 
25
  | Model | Average | AGIEval | GPT4All | TruthfulQA | Bigbench |
26
  |---|---:|---:|---:|---:|---:|
27
+ |[**Beagle14-7B**](https://huggingface.co/mlabonne/Beagle14-7B)| ****| ****| ****| ****| ****|
28
  | [mlabonne/Beagle14-7B](https://huggingface.co/mlabonne/Beagle14-7B) [πŸ“„](https://gist.github.com/mlabonne/f5a5bf8c0827bbec2f05b97cc62d642c) | 59.4 | 44.38 | 76.53 | 69.44 | 47.25 |
29
  | [mlabonne/NeuralDaredevil-7B](https://huggingface.co/mlabonne/NeuralDaredevil-7B) [πŸ“„](https://gist.github.com/mlabonne/cbeb077d1df71cb81c78f742f19f4155) | 59.39 | 45.23 | 76.2 | 67.61 | 48.52 |
30
  | [argilla/distilabeled-Marcoro14-7B-slerp](https://huggingface.co/argilla/distilabeled-Marcoro14-7B-slerp) [πŸ“„](https://gist.github.com/mlabonne/9082c4e59f4d3f3543c5eda3f4807040) | 58.93 | 45.38 | 76.48 | 65.68 | 48.18 |