Qwen 3 Collection Ampere's quantization formats (Q4_K_4 / Q8R16) require Ampere optimized llama.cpp available here: https://hub.docker.com/r/amperecomputingai/llama.cpp • 11 items • Updated 3 days ago • 2
Qwen 3 Collection Ampere's quantization formats (Q4_K_4 / Q8R16) require Ampere optimized llama.cpp available here: https://hub.docker.com/r/amperecomputingai/llama.cpp • 11 items • Updated 3 days ago • 2
Llama 3.2 Collection Ampere's quantization formats (Q4_K_4 / Q8R16) require Ampere optimized llama.cpp available here: https://hub.docker.com/r/amperecomputingai/llama.cpp • 2 items • Updated 3 days ago • 2
GPT-OSS Collection With gpt-oss models we recommend using native mxfp4 quantization. • 2 items • Updated 3 days ago • 1
Bielik v3.0 Collection Ampere's quantization formats (Q4_K_4 / Q8R16) require Ampere optimized llama.cpp available here: https://hub.docker.com/r/amperecomputingai/llama.cpp • 2 items • Updated 3 days ago • 1
Qwen 3 Collection Ampere's quantization formats (Q4_K_4 / Q8R16) require Ampere optimized llama.cpp available here: https://hub.docker.com/r/amperecomputingai/llama.cpp • 11 items • Updated 3 days ago • 2
GPT-OSS Collection With gpt-oss models we recommend using native mxfp4 quantization. • 2 items • Updated 3 days ago • 1
RakutenAI 7B Collection Ampere's quantization formats (Q4_K_4 / Q8R16) require Ampere optimized llama.cpp available here: https://hub.docker.com/r/amperecomputingai/llama.cpp • 3 items • Updated 3 days ago
RakutenAI 7B Collection Ampere's quantization formats (Q4_K_4 / Q8R16) require Ampere optimized llama.cpp available here: https://hub.docker.com/r/amperecomputingai/llama.cpp • 3 items • Updated 3 days ago