DXP-Zero-V1.0-24b-Small-Instruct-i1-GGUF

BF16 available here.

Recommended Settings

"temperature": 0.8,
"top_k": 40,
"top_p": 0.95,
"min_p": 0.05,
"repeat_last_n": 40,
"repeat_penalty": 1.2,

Run on Ollama

These are non-imatrix. I'll release the imatrix version later.

GGUF 3-bit Q3_K_M about 27 GB of vRAM/RAM:

ollama run hf.co/h34v7/DXP-Zero-V1.0-24b-Small-Instruct-i1-GGUF:Q3_K_M

GGUF 4-bit Q4_K_M about 30 GB of vRAM/RAM:

ollama run hf.co/h34v7/DXP-Zero-V1.0-24b-Small-Instruct-i1-GGUF:Q4_K_M

GGUF 5-bit Q5_K_M about 33 GB of vRAM/RAM:

ollama run hf.co/h34v7/DXP-Zero-V1.0-24b-Small-Instruct-i1-GGUF:Q5_K_M
Downloads last month
73
GGUF
Model size
23.6B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

3-bit

4-bit

5-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for h34v7/DXP-Zero-V1.0-24b-Small-Instruct-i1-GGUF

Quantized
(2)
this model

Collection including h34v7/DXP-Zero-V1.0-24b-Small-Instruct-i1-GGUF