|
--- |
|
license: mit |
|
language: |
|
- en |
|
pipeline_tag: text-generation |
|
library_name: transformers |
|
--- |
|
|
|
# π§ Model Card: Sam-2.5-4 (Pro Version) |
|
|
|
## π Overview |
|
|
|
**Sam-2.5-4** is the Pro continuation of the Sam-2.5 architecture series, designed for modular, multi-domain reasoning across math, dialogue, code, and open-domain tasks. It builds directly on **Sam-2.5-3**, continuing training for four additional epochs to deepen convergence, reduce domain bias, and improve generalization. |
|
|
|
This model is optimized for transparency, ablation-readiness, and deployment on both high-resource and low-resource devices (including Raspberry Pi). |
|
|
|
--- |
|
|
|
## 𧬠Model Lineage |
|
|
|
| Version | Description | |
|
|---------------|------------------------------------------------------------------------------| |
|
| **Sam-2.5-2** | GSM8K-heavy fine-tune; overfit to math; lacked domain balance | |
|
| **Sam-2.5-3** | Emergency patch; retrained from scratch on 4 datasets; balanced capabilities | |
|
| **Sam-2.5-4** | Pro version; continued training for 4 epochs; refined convergence and fluency| |
|
|
|
--- |
|
|
|
## π§ Architecture |
|
|
|
- Transformer-based, modular design |
|
- Registry-driven domain tagging and ablation toggles |
|
- Shape-adaptive loss functions with domain-aware diagnostics |
|
- Quantization-ready for Pi deployment |
|
- Verbose logging for batch-level feedback and anomaly tracing |
|
- Memory-safe serialization via `safetensors` |
|
|
|
--- |
|
|
|
## π Training Datasets |
|
|
|
| Dataset | Domain Focus | |
|
|------------------------|----------------------------------| |
|
| **GSM8K** | Mathematical reasoning | |
|
| **MultiWOZ** | Multi-turn dialogue & task flow | |
|
| **Alpaca-Code-Cleaned**| Code generation & logic | |
|
| **UltraChat-200k** | Open-domain conversation | |
|
|
|
- Datasets were concatenated, shuffled, and tagged for domain awareness |
|
- Replay and mixing strategies used to balance underrepresented domains |
|
- Training spanned **9 total epochs** (5 in -3, 4 in -4) |
|
|
|
--- |
|
|
|
## π Performance Summary |
|
|
|
| Metric | Value (Epoch 8β9) | |
|
|-------------------------|----------------------------------| |
|
| **Validation Loss** | β 2.95 (avg across domains) | |
|
| **Max Domain Loss** | < 3.4 (no domain exceeded) | |
|
| **Math Bias** | Resolved (loss spikes absorbed) | |
|
| **Dialogue Coherence** | Improved (MultiWOZ eval) | |
|
| **Code Determinism** | Increased (Alpaca eval) | |
|
| **Open-Domain Fluency** | Fewer hallucinations, better grounding | |
|
|
|
--- |
|
|
|
## π§ͺ Evaluation & Diagnostics |
|
|
|
- Loss spikes in early epochs traced to GSM8K; resolved by epoch 6 |
|
- Batch-level diagnostics printed per domain and token type |
|
- Attention stability improved on long-context prompts |
|
- Token transitions cleaner across dialogue and code tasks |
|
- Validation curve shows smooth convergence post-epoch 5 |
|
|
|
--- |
|
|
|
## π§© Deployment Notes |
|
|
|
- Compatible with Raspberry Pi (quantized + safetensors) |
|
- Supports CLI-based training diagnostics (loss, ETA, memory) |
|
- Registry hooks enable domain-specific ablation and extension |
|
- Ideal for benchmarking on GSM8K, MultiWOZ, UltraChat, and custom blends |
|
|
|
--- |
|
|
|
## π€ Intended Use |
|
|
|
- Research on modular Transformer architectures |
|
- Benchmarking across reasoning, dialogue, and code domains |
|
- Deployment on constrained hardware (e.g. Pi, ARM) |
|
- Community-driven extension and ablation testing |
|
|
|
--- |
|
|
|
## β οΈ Limitations |
|
|
|
- Still sensitive to prompt phrasing in edge cases |
|
- Long-context performance may degrade beyond 2k tokens |
|
- Requires domain tags for optimal generalization |
|
- Not trained on multimodal inputs (text-only) |
|
|
|
--- |
|
|
|
## π Acknowledgments |
|
|
|
Thanks to the open-source community, dataset curators, and contributors who helped shape Sam-2.5-4. This release reflects our shared commitment to transparent, inspectable, and extensible AI. |