reasoning-small-1B / README.md
Kameshr's picture
Update README.md
5b1774f verified
---
license: mit
language:
- en
tags:
- generated_from_trainer
- trl
- sft
- casual-lm
- text-generation
library_name: transformers
---
# Reasoning-Small-1B
A compact and efficient 1B parameter model fine-tuned for reasoning tasks using extracted Chain-of-Thought (CoT) data.
## Training Details
- **Base Model**: LLaMA 3.1 (1B parameters)
- **Dataset**: 2.67M extracted Chain-of-Thought examples
## Benchmarking
| Benchmark | Performance |
|-------------------|------------|
| GSM8K (Math) | 70.8% |
| ARC (Reasoning) | 69.4% |
| Math(CoT) | 40.8% |
## License
This repository is released under the MIT License.
---
library_name: transformers
tags:
- text-generation
- causal-lm
---