Uploaded model

  • Developed by: qingy2019
  • License: apache-2.0
  • Finetuned from model : unsloth/qwen2.5-14b-bnb-4bit

Huge thanks to Unsloth and the Huggingface TRL library.

This model is Qwen 2.5 14B fine tuned for a full epoch on the high quality garage-bAInd/Open-Platypus dataset for STEM reasoning.

Training Detail Value
Epochs 1
Steps 2077
Loss 0.4218
Batch size 4
Gradient Acc. Steps 3
Learning Rate 2e-4
LR Scheduler cosine
Rank 32
Rank-Stabilized LoRA Yes
Warm up steps 5
Weight Decay 0.01
Seed 3407

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 35.46
IFEval (0-Shot) 59.81
BBH (3-Shot) 47.75
MATH Lvl 5 (4-Shot) 23.11
GPQA (0-shot) 16.00
MuSR (0-shot) 17.95
MMLU-PRO (5-shot) 48.12
Downloads last month
77
Safetensors
Model size
14.8B params
Tensor type
FP16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for qingy2024/Qwen2.5-Math-14B-Instruct-Alpha

Merges
1 model
Quantizations
1 model

Collection including qingy2024/Qwen2.5-Math-14B-Instruct-Alpha

Evaluation results