saishshinde15's picture
Improve language tag (#1)
2183281 verified
metadata
base_model:
  - Qwen/Qwen2.5-3B-Instruct
tags:
  - gguf
  - q4
  - text-generation-inference
  - transformers
  - qwen2
  - trl
  - grpo
license: apache-2.0
language:
  - zho
  - eng
  - fra
  - spa
  - por
  - deu
  - ita
  - rus
  - jpn
  - kor
  - vie
  - tha
  - ara

TBH.AI Base Reasoning (GGUF - Q4)

  • Developed by: TBH.AI
  • License: apache-2.0
  • Fine-tuned from: Qwen/Qwen2.5-3B-Instruct
  • GGUF Format: 4-bit quantized (Q4) for optimized inference

Model Description

TBH.AI Base Reasoning (GGUF - Q4) is a 4-bit GGUF quantized version of saishshinde15/TBH.AI_Base_Reasoning, a fine-tuned model based on Qwen 2.5. This version is designed for high-efficiency inference on CPU/GPU with minimal memory usage, making it ideal for on-device applications and low-latency AI systems.

Trained using GRPO (General Reinforcement with Policy Optimization), the model excels in self-reasoning, logical deduction, and structured problem-solving, comparable to DeepSeek-R1. The Q4 quantization ensures significantly lower memory requirements while maintaining strong reasoning performance.

Features

  • 4-bit Quantization (Q4 GGUF): Optimized for low-memory, high-speed inference on compatible backends.
  • Self-Reasoning AI: Can process complex queries autonomously, generating logical and structured responses.
  • GRPO Fine-Tuning: Uses policy optimization for improved logical consistency and step-by-step reasoning.
  • Efficient On-Device Deployment: Works seamlessly with llama.cpp, KoboldCpp, GPT4All, and ctransformers.
  • Ideal for Logical Tasks: Best suited for research, coding logic, structured Q&A, and decision-making applications.

Limitations

  • This Q4 GGUF version is inference-only and does not support additional fine-tuning.
  • Quantization may slightly reduce response accuracy compared to FP16/full-precision models.
  • Performance depends on the execution environment and GGUF-compatible runtime.

Usage

Use this prompt for more detailed and personalized results. This is the recommended prompt as the model was tuned on it.

You are a reasoning model made by researcher at TBH.AI and your role is to respond in the following format only and in detail :

<reasoning>
...
</reasoning>
<answer>
...
</answer>

Use this prompt for concise representation of answers.

SYSTEM_PROMPT = """
Respond in the following format:
<reasoning>
...
</reasoning>
<answer>
...
</answer>
"""