A newer version of this model is available: Qwen/Qwen2.5-0.5B

Brianpuze/Qwen2-0.5B-Q4_K_M-Q4_0-GGUF

This repo contains GGUF quantized versions of Qwen/Qwen2-0.5B using llama.cpp.

Quantized Versions:

Run with llama.cpp

llama-cli --hf-repo Brianpuze/Qwen2-0.5B-Q4_K_M-Q4_0-GGUF --hf-file qwen2-0.5b-q4_k_m.gguf -p

Downloads last month
4
GGUF
Model size
494M params
Architecture
qwen2
Hardware compatibility
Log In to view the estimation

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Brianpuze/Qwen2-0.5B-Q4_K_M-Q4_0-GGUF

Base model

Qwen/Qwen2-0.5B
Quantized
(22)
this model