Missmatch between SmolLM2-360M-intermediate-checkpoints and SmolLM2-360M performance

#9
by Tobi-r9 - opened

Hey, I have run some evaluations (using lm-evaluation-harness). I wanted to understand the dynamics on some reasoning tasks. For that, I evaluated all checkpoints you provided at https://huggingface.co/HuggingFaceTB/SmolLM2-360M-intermediate-checkpoints. To my understanding, the final 2400k model from those checkpoints should equal the HuggingFaceTB/SmolLM2-360M model. However, I found a significant difference (see plot below). Could you clarify whether these are the correct checkpoints or if the HuggingFaceTB/SmolLM2-360M is a fine-tuned model?
(All evaluations were done with the same script.)

image.png

Sign up or log in to comment