Llama-3.2-Kapusta-3B-v8 GGUF Quantizations πŸ—²

Small and useful.

KapustaLogo256.png

This model was converted to GGUF format using llama.cpp.

For more information of the model, see the original model card: Khetterman/Llama-3.2-Kapusta-3B-v8.

Available Quantizations (β—•β€Ώβ—•)

My thanks to the authors of the original models, your work is incredible. Have a good time πŸ–€

Downloads last month
22
GGUF
Model size
3.61B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

4-bit

6-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for Khetterman/Llama-3.2-Kapusta-3B-v8-GGUF

Quantized
(6)
this model