oumi logo Made with Oumi

Documentation Blog Discord

oumi-ai/MiniMath-R1-1.5B

Supervised fine-tune of DeepSeek-R1-Distill-Qwen-1.5B using oumi-ai/MetaMathQA-R1.

Achieves 44.4% accuracy on MMLU-Pro-Math, the highest of any model with <=1.5B parameters.

Improves the base model's accuracy by +6 points.

Uses

Use as a conversational assistant for solving math problems with an exposed thought process.

Out-of-Scope Use

Smaller LLMs have limited capabilities and should be used with caution. Avoid using this model for purposes outside of mathematics.

Bias, Risks, and Limitations

This model was finetuned with DeepSeek-R1 data on top of an R1-distill model, so any biases or risks associated with those models may be present.

Training Details

Training Data

Training data: oumi-ai/MetaMathQA-R1

Training Procedure

Training notebook: Fine-Tuning Notebook

Evaluation

Environmental Impact

  • Hardware Type: H100
  • Hours used: 0.8 (0.1 * 8 GPUs)
  • Cloud Provider: Google Cloud Platform
  • Compute Region: us-east5
  • Carbon Emitted: 0.07 kg

Citation

@misc{miniMathR1_2025,
  author = {Jeremiah Greer},
  title = {MiniMath-R1-1.5B},
  month = {February},
  year = {2025},
  url = {https://huggingface.co/oumi-ai/MiniMath-R1-1.5B}
}

@software{oumi2025,
  author = {Oumi Community},
  title = {Oumi: an Open, End-to-end Platform for Building Large Foundation Models},
  month = {January},
  year = {2025},
  url = {https://github.com/oumi-ai/oumi}
}
Downloads last month
10
Safetensors
Model size
1.54B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for oumi-ai/MiniMath-R1-1.5B

Finetuned
(120)
this model

Dataset used to train oumi-ai/MiniMath-R1-1.5B