Fizzarolli commited on
Commit
fa6d905
·
verified ·
1 Parent(s): 1dfeef7

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +45 -0
README.md ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ - zh
6
+ base_model:
7
+ - mistralai/Mistral-Small-24B-Base-2501
8
+ tags:
9
+ - axolotl
10
+ ---
11
+ # Sisyphus 24b
12
+ Hundreds of dollars later.
13
+ Dozens of failed finetunes.
14
+ Sisyphus has balanced his rock on the summit.
15
+ One must have imagined him happy while pushing. Now, he is ecstatic.
16
+
17
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/634262af8d8089ebaefd410e/tHX80XuG_5HiW6F2hvvLe.jpeg)
18
+
19
+ ## About
20
+
21
+ This is a pretty generic finetune of the 24b base model for multiturn instruct. It's pretty coherent across a range of temps, assuming you use something like min-p or top-p. It also supports reasoning blocks.
22
+
23
+ ## System Prompts
24
+
25
+ I tested with the following Claude-like system prompts, however they were not trained in and any similar prompts can likely be used:
26
+
27
+ ### Non-Reasoning
28
+ ```
29
+ You are Claude, a helpful and harmless AI assistant created by Anthropic.
30
+ ```
31
+
32
+ ### Reasoning
33
+ ```
34
+ You are Claude, a helpful and harmless AI assistant created by Anthropic. Please contain all your thoughts in <think> </think> tags, and your final response right after the closing </think> tag.
35
+ ```
36
+
37
+ For reasoning, it's recommended to force the thinking (by prefilling `<think>\n` on the newest assistant response), as well as not including previous thought blocks in new requests.
38
+
39
+ ## Instruct Template
40
+
41
+ v7-Tekken, same as the original instruct model.
42
+
43
+ ## Dataset
44
+
45
+ This model was trained on [allura-org/inkstructmix-v0.1](https://hf.co/datasets/allura-org/inkstructmix-v0.1).