|
|
--- |
|
|
license: mit |
|
|
language: |
|
|
- en |
|
|
tags: |
|
|
- generated_from_trainer |
|
|
- trl |
|
|
- sft |
|
|
- casual-lm |
|
|
- text-generation |
|
|
library_name: transformers |
|
|
--- |
|
|
|
|
|
# Reasoning-Small-1B |
|
|
|
|
|
A compact and efficient 1B parameter model fine-tuned for reasoning tasks using extracted Chain-of-Thought (CoT) data. |
|
|
|
|
|
## Training Details |
|
|
- **Base Model**: LLaMA 3.1 (1B parameters) |
|
|
- **Dataset**: 2.67M extracted Chain-of-Thought examples |
|
|
|
|
|
|
|
|
## Benchmarking |
|
|
| Benchmark | Performance | |
|
|
|-------------------|------------| |
|
|
| GSM8K (Math) | 70.8% | |
|
|
| ARC (Reasoning) | 69.4% | |
|
|
| Math(CoT) | 40.8% | |
|
|
|
|
|
## License |
|
|
This repository is released under the MIT License. |
|
|
|
|
|
--- |
|
|
library_name: transformers |
|
|
tags: |
|
|
- text-generation |
|
|
- causal-lm |
|
|
--- |