Mistral-Small-24B-Base-2501-GGUF

This repo provides two GGUF quantizations of mistralai/Mistral-Small-24B-Base-2501:

Filename File size Description TLDR
Mistral-Small-24B-Base-2501-q8_0-q4_K_S.gguf 14.05GB q4_K_S quantization using q8_0 for token embeddings and output tensors Good quality, smaller size
Mistral-Small-24B-Base-2501-q8_0-q6_K.gguf 19.67GB q6_K quantization using q8_0 for token embeddings and output tensors Practically perfect quality, larger size
Downloads last month
10
GGUF
Model size
23.6B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

4-bit

6-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for ddh0/Mistral-Small-24B-Base-2501-GGUF

Quantized
(66)
this model