File size: 2,278 Bytes
03eba8a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
---
license: apache-2.0
language:
- en
base_model:
- prithivMLmods/Capricornus-MoT-1.7B-Supreme1
pipeline_tag: text-generation
library_name: transformers
tags:
- text-generation-inference
- math
- code
- science
---

# **Capricornus-MoT-1.7B-Supreme1-GGUF**

> **Capricornus-MoT-1.7B-Supreme1** is a **high-precision, multi-domain expert model** fine-tuned from **Qwen3-1.7B**, built for **code generation**, **mathematical reasoning**, **scientific analysis**, and **open technical inference**. Trained on the **Mixture of Thoughts (MoT)** dataset with combined expert clusters in **code, math, and science**, and enhanced with an **Open Code Reasoning** dataset, it delivers powerful symbolic and structured outputs in a wide range of STEM and reasoning domains.

## Model File

| File Name                                        | Size   | Format        | Description                              |
|--------------------------------------------------|--------|---------------|------------------------------------------|
| Capricornus-MoT-1.7B-Supreme1.BF16.gguf         | 3.45 GB| GGUF (BF16)   | BFloat16 precision model file            |
| Capricornus-MoT-1.7B-Supreme1.F16.gguf          | 3.45 GB| GGUF (F16)    | Float16 precision model file             |
| Capricornus-MoT-1.7B-Supreme1.F32.gguf          | 6.89 GB| GGUF (F32)    | Float32 precision model file             |
| Capricornus-MoT-1.7B-Supreme1.Q4_K_M.gguf       | 1.11 GB| GGUF (Q4_K_M) | 4-bit quantized model file               |
| Capricornus-MoT-1.7B-Supreme1.Q5_K_M.gguf       | 1.26 GB| GGUF (Q5_K_M) | 5-bit quantized model file               |
| Capricornus-MoT-1.7B-Supreme1.Q8_0.gguf         | 1.83 GB| GGUF (Q8_0)   | 8-bit quantized model file               |
| config.json                                     | 31 B   | JSON          | Configuration file                       |
| .gitattributes                                  | 1.98 kB| Text          | Git attributes configuration             |

## Quants Usage 

(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)

Here is a handy graph by ikawrakow comparing some lower-quality quant
types (lower is better):

![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png)