Improve model card: Add paper URL, pipeline tag and Github URL
#6
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,14 +1,15 @@
|
|
| 1 |
---
|
| 2 |
-
|
|
|
|
|
|
|
| 3 |
datasets:
|
| 4 |
- BAAI/TACO
|
| 5 |
- tasksource/PRM800K
|
| 6 |
language:
|
| 7 |
- en
|
| 8 |
-
|
| 9 |
-
- Qwen/Qwen2.5-32B-Instruct
|
| 10 |
-
- NovaSky-AI/Sky-T1-32B-Preview
|
| 11 |
license: apache-2.0
|
|
|
|
| 12 |
---
|
| 13 |
|
| 14 |
## Model Details
|
|
@@ -18,9 +19,11 @@ license: apache-2.0
|
|
| 18 |
<!-- Provide a longer summary of what this model is. -->
|
| 19 |
|
| 20 |
This is a 32B reasoning model preference optimized on top of Sky-T1-32B-Preview to significantly reduce generation lengths while maintaining accuracy. The performance is on par with o1-preview model in both math and coding, while reducing generation lengths by up to 57% relative to Sky-T1-32B-Preview.
|
| 21 |
-
Please see our [blog post](https://novasky-ai.github.io/posts/reduce-overthinking/) for more details.
|
| 22 |
|
| 23 |
- **Developed by:** NovaSky Team from Sky Computing Lab at UC Berkeley.
|
|
|
|
|
|
|
| 24 |
|
| 25 |
## Training Details
|
| 26 |
|
|
@@ -71,3 +74,4 @@ Please considering citing our blog post if you found it useful for your research
|
|
| 71 |
note = {Accessed: 2025-01-23},
|
| 72 |
year = {2025}
|
| 73 |
}
|
|
|
|
|
|
| 1 |
---
|
| 2 |
+
base_model:
|
| 3 |
+
- Qwen/Qwen2.5-32B-Instruct
|
| 4 |
+
- NovaSky-AI/Sky-T1-32B-Preview
|
| 5 |
datasets:
|
| 6 |
- BAAI/TACO
|
| 7 |
- tasksource/PRM800K
|
| 8 |
language:
|
| 9 |
- en
|
| 10 |
+
library_name: transformers
|
|
|
|
|
|
|
| 11 |
license: apache-2.0
|
| 12 |
+
pipeline_tag: text-generation
|
| 13 |
---
|
| 14 |
|
| 15 |
## Model Details
|
|
|
|
| 19 |
<!-- Provide a longer summary of what this model is. -->
|
| 20 |
|
| 21 |
This is a 32B reasoning model preference optimized on top of Sky-T1-32B-Preview to significantly reduce generation lengths while maintaining accuracy. The performance is on par with o1-preview model in both math and coding, while reducing generation lengths by up to 57% relative to Sky-T1-32B-Preview.
|
| 22 |
+
Please see our [blog post](https://novasky-ai.github.io/posts/reduce-overthinking/) and [Sky-T1 blog post](https://novasky-ai.github.io/posts/sky-t1/) for more details.
|
| 23 |
|
| 24 |
- **Developed by:** NovaSky Team from Sky Computing Lab at UC Berkeley.
|
| 25 |
+
- **Paper:** [LLMs Can Easily Learn to Reason from Demonstrations Structure, not content, is what matters!](https://hf.co/papers/2502.07374)
|
| 26 |
+
- **Code:** [https://github.com/NovaSky-AI/SkyThought](https://github.com/NovaSky-AI/SkyThought)
|
| 27 |
|
| 28 |
## Training Details
|
| 29 |
|
|
|
|
| 74 |
note = {Accessed: 2025-01-23},
|
| 75 |
year = {2025}
|
| 76 |
}
|
| 77 |
+
```
|