---
base_model:
- Qwen/Qwen2.5-3B-Instruct
tags:
- gguf
- q4
- text-generation-inference
- transformers
- qwen2
- trl
- grpo
license: apache-2.0
language:
- zho
- eng
- fra
- spa
- por
- deu
- ita
- rus
- jpn
- kor
- vie
- tha
- ara
---
# TBH.AI Base Reasoning (GGUF - Q4)
- **Developed by:** TBH.AI
- **License:** apache-2.0
- **Fine-tuned from:** Qwen/Qwen2.5-3B-Instruct
- **GGUF Format:** 4-bit quantized (Q4) for optimized inference
## **Model Description**
TBH.AI Base Reasoning (GGUF - Q4) is a **4-bit GGUF quantized** version of `saishshinde15/TBH.AI_Base_Reasoning`, a fine-tuned model based on **Qwen 2.5**. This version is designed for **high-efficiency inference on CPU/GPU with minimal memory usage**, making it ideal for on-device applications and low-latency AI systems.
Trained using **GRPO (General Reinforcement with Policy Optimization)**, the model excels in **self-reasoning, logical deduction, and structured problem-solving**, comparable to **DeepSeek-R1**. The **Q4 quantization** ensures significantly lower memory requirements while maintaining strong reasoning performance.
## **Features**
- **4-bit Quantization (Q4 GGUF):** Optimized for low-memory, high-speed inference on compatible backends.
- **Self-Reasoning AI:** Can process complex queries autonomously, generating logical and structured responses.
- **GRPO Fine-Tuning:** Uses policy optimization for improved logical consistency and step-by-step reasoning.
- **Efficient On-Device Deployment:** Works seamlessly with **llama.cpp, KoboldCpp, GPT4All, and ctransformers**.
- **Ideal for Logical Tasks:** Best suited for **research, coding logic, structured Q&A, and decision-making applications**.
## **Limitations**
- This **Q4 GGUF version is inference-only** and does not support additional fine-tuning.
- Quantization may slightly reduce response accuracy compared to FP16/full-precision models.
- Performance depends on the execution environment and GGUF-compatible runtime.
## **Usage**
# Use this prompt for more detailed and personalized results. This is the recommended prompt as the model was tuned on it.
```python
You are a reasoning model made by researcher at TBH.AI and your role is to respond in the following format only and in detail :
...
...
```
# Use this prompt for concise representation of answers.
```python
SYSTEM_PROMPT = """
Respond in the following format:
...
...
"""