The model was trained on paired preferences from the MathDial and MRBench datasets.

To find more information and to cite, see:

@article{macina2025mathtutorbench,
      title={MathTutorBench: A Benchmark for Measuring Open-ended\\ Pedagogical Capabilities of LLM Tutors}, 
      author={Jakub Macina, Nico Daheim, Ido Hakimi, Manu Kapur, Iryna Gurevych, Mrinmaya Sachan},
      year={2025},
      eprint={2502.18940},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2502.18940},
}
Downloads last month
31
Safetensors
Model size
1.54B params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for eth-nlped/Qwen2.5-1.5B-pedagogical-rewardmodel

Base model

Qwen/Qwen2.5-1.5B
Finetuned
(379)
this model

Dataset used to train eth-nlped/Qwen2.5-1.5B-pedagogical-rewardmodel