GGUF
llama.cpp
polaris
ollama
dzur658's picture
Update README.md
4410f3e verified
metadata
license: apache-2.0
datasets:
  - POLARIS-Project/Polaris-Dataset-53K
base_model:
  - POLARIS-Project/Polaris-4B-Preview
library_name: llama.cpp
tags:
  - polaris
  - gguf
  - ollama

Polaris GGUF Quants

This repository contains GGUF (GPT-Generated Unified Format) model files for the Polaris model.

These files were quantized using an importance matrix generated from the Polaris dataset for optimal performance. The imatrix file can be viewed on the Github for this repository (coming soon). The importance matrix was generated using 10k semi-randomly chosen examples from the Polaris Dataset

Associated Github

The Github for this project contains the required modelfile settings necessary to run this model with Ollama along with some of the scripts I used to create these imatrix quantizations.

Accreditation

All credit for the original Polaris model belongs to their team and their affiliated organizations. Thank you for all of your hard work, and open publishing of the models and research 😊, the POLARIS recipe will be absolutely crucial for edge LLM computing and if you haven't read their original paper I'd reccomend giving it a good read to truly appreciate it.

@misc{Polaris2025,
    title = {POLARIS: A Post-Training Recipe for Scaling Reinforcement Learning on Advanced Reasoning Models},
    url = {https://hkunlp.github.io/blog/2025/Polaris},
    author = {An, Chenxin and Xie, Zhihui and Li, Xiaonan and Li, Lei and Zhang, Jun and Gong, Shansan and Zhong, Ming and Xu, Jingjing and Qiu, Xipeng and Wang, Mingxuan and Kong, Lingpeng}
    year = {2025}
}

Ollama Integration

To use these models with Ollama, llmstudio, llama.cpp, etc. select your desired quantization level from the dropdown menu (e.g., Q4_K_M) and use the provided command.

Note: Replace Q4_K_M with the tag for the specific model version you wish to download.

ollama pull hf.co/dzur658/Polaris-4B-Preview-IQ-GGUF:Q4_K_M