mlabonne commited on
Commit
2ed062e
1 Parent(s): 1294f9e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +28 -2
README.md CHANGED
@@ -20,9 +20,35 @@ base_model:
20
 
21
  Meta-Llama-3-120B-Instruct is a self-merge with [meta-llama/Meta-Llama-3-70B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct).
22
 
23
- It was inspired by large merges like [alpindale/goliath-120b](https://huggingface.co/alpindale/goliath-120b), [nsfwthrowitaway69/Venus-120b-v1.0](https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.0), [cognitivecomputations/MegaDolphin-120b](https://huggingface.co/cognitivecomputations/MegaDolphin-120b), and [wolfram/miquliz-120b-v2.0](https://huggingface.co/wolfram/miquliz-120b-v2.0).
24
 
25
- No eval yet, but it is approved by Eric Hartford: https://twitter.com/erhartford/status/1787050962114207886
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
 
27
  ## 🧩 Configuration
28
 
 
20
 
21
  Meta-Llama-3-120B-Instruct is a self-merge with [meta-llama/Meta-Llama-3-70B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct).
22
 
23
+ It was inspired by large merges like:
24
 
25
+ - [alpindale/goliath-120b](https://huggingface.co/alpindale/goliath-120b)
26
+ - [nsfwthrowitaway69/Venus-120b-v1.0](https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.0)
27
+ - [cognitivecomputations/MegaDolphin-120b](https://huggingface.co/cognitivecomputations/MegaDolphin-120b)
28
+ - [wolfram/miquliz-120b-v2.0](https://huggingface.co/wolfram/miquliz-120b-v2.0).
29
+
30
+ ## 🔍 Applications
31
+
32
+ I recommend using this model for creative writing. It uses the Llama 3 chat template with a default context window of 8K (can be extended with rope theta).
33
+
34
+ Check the examples in the evaluation section to get an idea of its performance.
35
+
36
+ ## ⚡ Quantized models
37
+
38
+ Thanks to [Eric Hartford](https://huggingface.co/ehartford), [elinas](https://huggingface.co/elinas), and the [mlx-community](https://huggingface.co/mlx-community) for providing these models.
39
+
40
+ * **GGUF**: https://huggingface.co/cognitivecomputations/Meta-Llama-3-120B-Instruct-gguf
41
+ * **EXL2**: https://huggingface.co/elinas/Meta-Llama-3-120B-Instruct-4.0bpw-exl2
42
+ * **mlx**: https://huggingface.co/mlx-community/Meta-Llama-3-120B-Instruct-4bit
43
+
44
+ ## 🏆 Evaluation
45
+
46
+ The model looks excellent for creating writing tasks, outperforming GPT-4. Thanks again to [Eric Hartford](https://huggingface.co/ehartford) for noticing this.
47
+
48
+ * **X thread by Eric Hartford (creative writing)**: https://twitter.com/erhartford/status/1787050962114207886
49
+ * **X thread by Daniel Kaiser (creative wrirting)**: https://twitter.com/spectate_or/status/1787257261309518101
50
+ * **X thread by Simon (reasoning)**: https://twitter.com/NewDigitalEdu/status/1787403266894020893
51
+ * **r/LocalLLaMa**: https://www.reddit.com/r/LocalLLaMA/comments/1cl525q/goliath_lovers_where_is_the_feedback_about/
52
 
53
  ## 🧩 Configuration
54