Update README.md
Browse files
README.md
CHANGED
@@ -50,7 +50,7 @@ Advanced, high-quality and an easy to run reasoning for a small size that you ca
|
|
50 |
|
51 |
At its original quality, it runs at ~250 tokens/second on a single Friendli H100 Nvidia GPU.
|
52 |
|
53 |
-
Trained similarly to Deepseek R1, we used Phi-3.5-mini as a base model, then we've SFT fine tuned on reasoning using our own private superthoughts instruct dataset which includes a mix of code, website generation, day-to-day chats, math and counting problems & summerization. after the SFT fine tuning we used GRPO to further amplify it's mathematics & problem solving abilities.
|
54 |
Unlike the LITE version of superthoughts, we've fixed a few issues in our instruction/sft dataset, and changed the GRPO code so the model would be more significantly conversational (eg; if you ask it "Whats the weather like?" the lite version would think and probably just answer with "good", unlike superthoughts mini which would start an actual conversation). This model has very strong reasoning abillities yet does not over-think. We personally have found that at least in GSM8K, over thinking did not help much, rather it just made the model get confused and waste a lot of tokens. so due to this, superthoughts mini does not usually over-think.
|
55 |
|
56 |
# Format & Examples
|
|
|
50 |
|
51 |
At its original quality, it runs at ~250 tokens/second on a single Friendli H100 Nvidia GPU.
|
52 |
|
53 |
+
Trained similarly to Deepseek R1, we used Phi-3.5-mini (and it uses the same chat template as phi-3) as a base model, then we've SFT fine tuned on reasoning using our own private superthoughts instruct dataset which includes a mix of code, website generation, day-to-day chats, math and counting problems & summerization. after the SFT fine tuning we used GRPO to further amplify it's mathematics & problem solving abilities.
|
54 |
Unlike the LITE version of superthoughts, we've fixed a few issues in our instruction/sft dataset, and changed the GRPO code so the model would be more significantly conversational (eg; if you ask it "Whats the weather like?" the lite version would think and probably just answer with "good", unlike superthoughts mini which would start an actual conversation). This model has very strong reasoning abillities yet does not over-think. We personally have found that at least in GSM8K, over thinking did not help much, rather it just made the model get confused and waste a lot of tokens. so due to this, superthoughts mini does not usually over-think.
|
55 |
|
56 |
# Format & Examples
|