IntelligentEstate/miniclaus-qw1.5B-UNAMGS-GRPO-Q8_0-GGUF a Thinking or Brooding Model
This model Is showing impressive speed and accuracy with situational and complex math reasoning the Group Realative Policy Optimization applied has shown to smooth out some of the overall issues while adding accuracy in the small model when quantized.(So much so the process inspire the creation of berkley's DeepScaleR-1) Showing improvements in my favorite little model and giving it a bit more cohearance and in many areas overtaking Qwen's 2.5 3B original model. A VERY impressive Model by FBLGIT. and out of 10 different Quants this one seems to be the best in local intel based testing.
Still the best little guy for it's size, THANKS for the present FBLGIT/miniclaus-qw1.5B-UNAMGS-GRPO-Q8_0-GGUF
This model was converted to GGUF format from fblgit/miniclaus-qw1.5B-UNAMGS-GRPO
using llama.cpp via the ggml.ai's GGUF-my-repo space.
Refer to the original model card for more details on the model.
Use with llama.cpp
Install llama.cpp through brew (works on Mac and Linux)
brew install llama.cpp
Invoke the llama.cpp server or the CLI.
GPT4ALL/Ollama: use standard qwen templates/prompting opening context window for length may applie qwen "Reasoner" template for analysing function.
- Downloads last month
- 12
Model tree for IntelligentEstate/miniclaus-qw1.5B-UNAMGS-GRPO-Q8_0-GGUF
Base model
Qwen/Qwen2.5-1.5B