IntelligentEstate/miniclaus-qw1.5B-UNAMGS-GRPO-Q8_0-GGUF a Thinking or Brooding Model

This model Is showing impressive speed and accuracy with situational and complex math reasoning the Group Realative Policy Optimization applied has shown to smooth out some of the overall issues while adding accuracy in the small model when quantized.(So much so the process inspire the creation of berkley's DeepScaleR-1) Showing improvements in my favorite little model and giving it a bit more cohearance and in many areas overtaking Qwen's 2.5 3B original model. A VERY impressive Model by FBLGIT. and out of 10 different Quants this one seems to be the best in local intel based testing.

Screenshot 2024-12-18 at 09-00-31 Ideogram.png

Still the best little guy for it's size, THANKS for the present FBLGIT/miniclaus-qw1.5B-UNAMGS-GRPO-Q8_0-GGUF

This model was converted to GGUF format from fblgit/miniclaus-qw1.5B-UNAMGS-GRPO using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model.

Use with llama.cpp

Install llama.cpp through brew (works on Mac and Linux)

brew install llama.cpp

Invoke the llama.cpp server or the CLI.

GPT4ALL/Ollama: use standard qwen templates/prompting opening context window for length may applie qwen "Reasoner" template for analysing function.

Downloads last month
12
GGUF
Model size
1.54B params
Architecture
qwen2

8-bit

Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no pipeline_tag.

Model tree for IntelligentEstate/miniclaus-qw1.5B-UNAMGS-GRPO-Q8_0-GGUF

Dataset used to train IntelligentEstate/miniclaus-qw1.5B-UNAMGS-GRPO-Q8_0-GGUF