reasoning-small-1B / README.md
Kameshr's picture
Update README.md
5b1774f verified
metadata
license: mit
language:
  - en
tags:
  - generated_from_trainer
  - trl
  - sft
  - casual-lm
  - text-generation
library_name: transformers

Reasoning-Small-1B

A compact and efficient 1B parameter model fine-tuned for reasoning tasks using extracted Chain-of-Thought (CoT) data.

Training Details

  • Base Model: LLaMA 3.1 (1B parameters)
  • Dataset: 2.67M extracted Chain-of-Thought examples

Benchmarking

Benchmark Performance
GSM8K (Math) 70.8%
ARC (Reasoning) 69.4%
Math(CoT) 40.8%

License

This repository is released under the MIT License.


library_name: transformers tags: - text-generation - causal-lm