This is a reasoning model that almost always shows the <think>
prefix, even outside of RP. It was a quick fine-tune done just for fun.
It works terribly in languages other than English.
Don't evaluate this as something serious at the moment.
Training Details
- Sequence Length: 8192
- Epochs: 1 epoch
- Full fine-tuning
- Learning Rate: 0.00008
- Scheduler: Cosine
- Total batch size (4 x 16 x 1) = 64
- Downloads last month
- 16
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support