🧠 Model Card: Sam-2.5-2

Overview

Sam-2.5-2 is a fine-tuned variant of Sam2.5, optimized for chain-of-thought reasoning on GSM8K. It retains modular, ablation-ready architecture and demonstrates strong generalization across arithmetic and logic-heavy prompts.


πŸ”§ Architecture

Component Value
Base Model Sam2.5
Layers Unchanged
Heads Unchanged
FF Multiplier Unchanged
Dropout Unchanged
Tokenizer AutoTokenizer
Shared Weights lm_head ↔ embed (cloned during save)

πŸ§ͺ Training Details

Parameter Value
Dataset GSM8K
Epochs 2
Batch Size 2
Max Length 512
Optimizer AdamW
Learning Rate 1e-4
Replay Mixing None
Early Stopping Manual checkpointing

πŸ“‰ Performance Metrics

Metric Epoch 1 Epoch 2
Final Train Loss 0.7826 2.7956
Validation Loss 2.5932 1.8989
Perplexity 13.37 6.68

πŸ” Output Quality

  • βœ… Fluent chain-of-thought steps
  • βœ… Accurate arithmetic reasoning
  • βœ… Consistent use of scratchpad format (<<...>>)
  • βœ… Stable token alignment across nested logic

πŸ’Ύ Checkpointing

  • Safe save logic applied to avoid shared memory errors
  • Format: .safetensors
  • Best model: checkpoints/epoch_2_loss_1.8989/
  • Final model: checkpoints/final/
Downloads last month
16
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support