๐Ÿง  Qwen-0.6B Reasoning โ€“ XformAI Fine-Tuned Model

Model: XformAI-india/qwen-0.6b-reasoning
Base Model: Qwen/Qwen3-0.6B
Architecture: Transformer decoder (GPT-style)
Fine-Tuned By: XformAI
Release Date: May 2025
License: MIT


๐Ÿง  What is it?

qwen-0.6b-reasoning is a compact transformer model fine-tuned for reasoning, logic, and analytical thinking.

Despite its size, it demonstrates strong performance across:

  • ๐Ÿงฉ Riddles & Puzzles
  • ๐Ÿงฎ Math Word Problems
  • ๐Ÿง  Symbolic Reasoning
  • ๐Ÿ’ฌ Chain-of-Thought Prompting
  • ๐Ÿ” Common Sense Logic

Fine-tuned on a curated instruction-style dataset focused on multi-step reasoning.


๐Ÿš€ Why it Matters

  • Performs like a 7B model on reasoning benchmarks
  • Lightweight (600M) and can run on CPU or mobile edge devices
  • Excels in step-by-step explanations and problem solving

๐Ÿงช Fine-Tuning Overview


Category Detail
Base Model Qwen 0.6B
Target Objective Reasoning, logic, CoT
Fine-Tuning Type Instruction
Optimizer AdamW (LoRA tuning)
Precision bfloat16
Epochs 2
Max Tokens 2048

๐Ÿงฉ Prompt Example

from transformers import AutoTokenizer, AutoModelForCausalLM

model = AutoModelForCausalLM.from_pretrained("XformAI-india/qwen-0.6b-reasoning")
tokenizer = AutoTokenizer.from_pretrained("XformAI-india/qwen-0.6b-reasoning")

prompt = "A farmer has 17 sheep. All but 9 run away. How many are left?"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Downloads last month
6
Safetensors
Model size
596M params
Tensor type
FP16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for XformAI-india/qwen-0.6b-reasoning

Finetuned
Qwen/Qwen3-0.6B
Finetuned
(124)
this model

Dataset used to train XformAI-india/qwen-0.6b-reasoning