Solshine commited on
Commit
83f820d
·
verified ·
1 Parent(s): c9d33e6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +42 -1
README.md CHANGED
@@ -2,6 +2,7 @@
2
  base_model: Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit
3
  datasets:
4
  - mahiatlinux/Reflection-Dataset-v2
 
5
  language:
6
  - en
7
  license: llama3.1
@@ -21,6 +22,46 @@ tags:
21
  This model was converted to GGUF format from [`Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit`](https://huggingface.co/Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
22
  Refer to the [original model card](https://huggingface.co/Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit) for more details on the model.
23
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  ## Use with llama.cpp
25
  Install llama.cpp through brew (works on Mac and Linux)
26
 
@@ -59,4 +100,4 @@ Step 3: Run inference through the main binary.
59
  or
60
  ```
61
  ./llama-server --hf-repo Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit-Q4_K_M-GGUF --hf-file reflection-llama-3.1-8b-solshine-trainround3-16bit-q4_k_m.gguf -c 2048
62
- ```
 
2
  base_model: Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit
3
  datasets:
4
  - mahiatlinux/Reflection-Dataset-v2
5
+ - Harshkmr/orca-math-word-reflection
6
  language:
7
  - en
8
  license: llama3.1
 
22
  This model was converted to GGUF format from [`Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit`](https://huggingface.co/Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
23
  Refer to the [original model card](https://huggingface.co/Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit) for more details on the model.
24
 
25
+ # Uploaded model
26
+
27
+ - **Developed by:** Solshine (Caleb DeLeeuw)
28
+ - **License:** LLama 3.1 License
29
+ - **Finetuned from model :** Solshine/reflection-llama-3.1-8B-Solshine-trainround2-16bit
30
+
31
+ Inspired by and featuring the Reflection Tuning technique pioneered by Matt Shumer (possibly earlier innovated by the team at Anthropic.)
32
+
33
+ *To the authors' knowledge, this is V3 of the first "reflection tuned" Llama 3.1 8B LLM*
34
+
35
+
36
+ **As per the inspiring model "mattshumer/Reflection-Llama-3.1-70B" (this mode was not used in the training process nor as a foundational model, but only served as inspiration) :**
37
+
38
+ '''
39
+
40
+ During sampling, the model will start by outputting reasoning inside <thinking> and </thinking> tags, and then once it is satisfied with its reasoning, it will output the final answer inside <output> and </output> tags. Each of these tags are special tokens, trained into the model.
41
+
42
+ This enables the model to separate its internal thoughts and reasoning from its final answer, improving the experience for the user.
43
+
44
+ Inside the <thinking> section, the model may output one or more <reflection> tags, which signals the model has caught an error in its reasoning and will attempt to correct it before providing a final answer.
45
+
46
+ System Prompt:
47
+ The system prompt used for training this model is:
48
+
49
+ You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside <thinking> tags, and then provide your final response inside <output> tags. If you detect that you made a mistake in your reasoning at any point, correct yourself inside <reflection> tags.
50
+
51
+ We recommend using this exact system prompt to get the best results from Reflection Llama-3.1 70B. You may also want to experiment combining this system prompt with your own custom instructions to customize the behavior of the model.
52
+
53
+ Chat Format:
54
+ As mentioned above, the model uses the standard Llama 3.1 chat format. Here’s an example:
55
+
56
+ <|begin_of_text|><|start_header_id|>system<|end_header_id|>
57
+
58
+ You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside <thinking> tags, and then provide your final response inside <output> tags. If you detect that you made a mistake in your reasoning at any point, correct yourself inside <reflection> tags.<|eot_id|><|start_header_id|>user<|end_header_id|>
59
+
60
+ what is 2+2?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
61
+
62
+ '''
63
+
64
+
65
  ## Use with llama.cpp
66
  Install llama.cpp through brew (works on Mac and Linux)
67
 
 
100
  or
101
  ```
102
  ./llama-server --hf-repo Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit-Q4_K_M-GGUF --hf-file reflection-llama-3.1-8b-solshine-trainround3-16bit-q4_k_m.gguf -c 2048
103
+ ```