ysn-rfd/HelpingAI2.5-5B-GGUF

This model was converted to GGUF format from HelpingAI/HelpingAI2.5-5B using llama.cpp via the ggml.ai's all-gguf-same-where space. Refer to the original model card for more details on the model.

βœ… Quantized Models Download List

✨ Recommended for CPU: Q4_K_M | ⚑ Recommended for ARM CPU: Q4_0 | πŸ† Best Quality: Q8_0

πŸš€ Download πŸ”’ Type πŸ“ Notes
Download Q2_K Basic quantization
Download Q3_K_S Small size
Download Q3_K_M Balanced quality
Download Q3_K_L Better quality
Download Q4_0 Fast on ARM
Download Q4_K_S Fast, recommended
Download Q4_K_M ⭐ Best balance
Download Q5_0 Good quality
Download Q5_K_S Balanced
Download Q5_K_M High quality
Download Q6_K πŸ† Very good quality
Download Q8_0 ⚑ Fast, best quality
Download F16 Maximum accuracy

πŸ’‘ Tip: Use F16 for maximum precision when quality is critical

Downloads last month
17
GGUF
Model size
5.13B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

4-bit

5-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for ysn-rfd/HelpingAI2.5-5B-GGUF

Quantized
(9)
this model