Improve model card: Add paper URL, pipeline tag and Github URL
#6
by
nielsr
HF staff
- opened
README.md
CHANGED
@@ -1,14 +1,15 @@
|
|
1 |
---
|
2 |
-
|
|
|
|
|
3 |
datasets:
|
4 |
- BAAI/TACO
|
5 |
- tasksource/PRM800K
|
6 |
language:
|
7 |
- en
|
8 |
-
|
9 |
-
- Qwen/Qwen2.5-32B-Instruct
|
10 |
-
- NovaSky-AI/Sky-T1-32B-Preview
|
11 |
license: apache-2.0
|
|
|
12 |
---
|
13 |
|
14 |
## Model Details
|
@@ -18,9 +19,11 @@ license: apache-2.0
|
|
18 |
<!-- Provide a longer summary of what this model is. -->
|
19 |
|
20 |
This is a 32B reasoning model preference optimized on top of Sky-T1-32B-Preview to significantly reduce generation lengths while maintaining accuracy. The performance is on par with o1-preview model in both math and coding, while reducing generation lengths by up to 57% relative to Sky-T1-32B-Preview.
|
21 |
-
Please see our [blog post](https://novasky-ai.github.io/posts/reduce-overthinking/) for more details.
|
22 |
|
23 |
- **Developed by:** NovaSky Team from Sky Computing Lab at UC Berkeley.
|
|
|
|
|
24 |
|
25 |
## Training Details
|
26 |
|
@@ -71,3 +74,4 @@ Please considering citing our blog post if you found it useful for your research
|
|
71 |
note = {Accessed: 2025-01-23},
|
72 |
year = {2025}
|
73 |
}
|
|
|
|
1 |
---
|
2 |
+
base_model:
|
3 |
+
- Qwen/Qwen2.5-32B-Instruct
|
4 |
+
- NovaSky-AI/Sky-T1-32B-Preview
|
5 |
datasets:
|
6 |
- BAAI/TACO
|
7 |
- tasksource/PRM800K
|
8 |
language:
|
9 |
- en
|
10 |
+
library_name: transformers
|
|
|
|
|
11 |
license: apache-2.0
|
12 |
+
pipeline_tag: text-generation
|
13 |
---
|
14 |
|
15 |
## Model Details
|
|
|
19 |
<!-- Provide a longer summary of what this model is. -->
|
20 |
|
21 |
This is a 32B reasoning model preference optimized on top of Sky-T1-32B-Preview to significantly reduce generation lengths while maintaining accuracy. The performance is on par with o1-preview model in both math and coding, while reducing generation lengths by up to 57% relative to Sky-T1-32B-Preview.
|
22 |
+
Please see our [blog post](https://novasky-ai.github.io/posts/reduce-overthinking/) and [Sky-T1 blog post](https://novasky-ai.github.io/posts/sky-t1/) for more details.
|
23 |
|
24 |
- **Developed by:** NovaSky Team from Sky Computing Lab at UC Berkeley.
|
25 |
+
- **Paper:** [LLMs Can Easily Learn to Reason from Demonstrations Structure, not content, is what matters!](https://hf.co/papers/2502.07374)
|
26 |
+
- **Code:** [https://github.com/NovaSky-AI/SkyThought](https://github.com/NovaSky-AI/SkyThought)
|
27 |
|
28 |
## Training Details
|
29 |
|
|
|
74 |
note = {Accessed: 2025-01-23},
|
75 |
year = {2025}
|
76 |
}
|
77 |
+
```
|