๐ง Qwen-0.6B Reasoning โ XformAI Fine-Tuned Model
Model: XformAI-india/qwen-0.6b-reasoning
Base Model: Qwen/Qwen3-0.6B
Architecture: Transformer decoder (GPT-style)
Fine-Tuned By: XformAI
Release Date: May 2025
License: MIT
๐ง What is it?
qwen-0.6b-reasoning
is a compact transformer model fine-tuned for reasoning, logic, and analytical thinking.
Despite its size, it demonstrates strong performance across:
- ๐งฉ Riddles & Puzzles
- ๐งฎ Math Word Problems
- ๐ง Symbolic Reasoning
- ๐ฌ Chain-of-Thought Prompting
- ๐ Common Sense Logic
Fine-tuned on a curated instruction-style dataset focused on multi-step reasoning.
๐ Why it Matters
- Performs like a 7B model on reasoning benchmarks
- Lightweight (600M) and can run on CPU or mobile edge devices
- Excels in step-by-step explanations and problem solving
๐งช Fine-Tuning Overview
Category | Detail |
---|---|
Base Model | Qwen 0.6B |
Target Objective | Reasoning, logic, CoT |
Fine-Tuning Type | Instruction |
Optimizer | AdamW (LoRA tuning) |
Precision | bfloat16 |
Epochs | 2 |
Max Tokens | 2048 |
๐งฉ Prompt Example
from transformers import AutoTokenizer, AutoModelForCausalLM
model = AutoModelForCausalLM.from_pretrained("XformAI-india/qwen-0.6b-reasoning")
tokenizer = AutoTokenizer.from_pretrained("XformAI-india/qwen-0.6b-reasoning")
prompt = "A farmer has 17 sheep. All but 9 run away. How many are left?"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
- Downloads last month
- 6
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support