L'ia is a Reasoning Model trained on 500 steps in Unsloth GRPO based on 49,909 (oops) datapoints of Lambda Calculus.
This active experiment is a test to see if a model can be trained to primarily think in pure simple logic. Maybe after more time we can make the Vulcans proud!
- Developed by: Creekside
- License: apache-2.0
- Finetuned from model : unsloth/qwen2.5-3b-instruct-unsloth-bnb-4bit
This qwen2 model was trained 2x faster with Unsloth and Huggingface's TRL library.
- Downloads last month
- 86
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.