matrixportal/X-Ray_Alpha-GGUF
This model was converted to GGUF format from SicariusSicariiStuff/X-Ray_Alpha
using llama.cpp via the ggml.ai's all-gguf-same-where space.
Refer to the original model card for more details on the model.
β Quantized Models Download List
π Recommended Quantizations
- β¨ General CPU Use:
Q4_K_M
(Best balance of speed/quality) - π± ARM Devices:
Q4_0
(Optimized for ARM CPUs) - π Maximum Quality:
Q8_0
(Near-original quality)
π¦ Full Quantization Options
π Download | π’ Type | π Notes |
---|---|---|
Download | Basic quantization | |
Download | Small size | |
Download | Balanced quality | |
Download | Better quality | |
Download | Fast on ARM | |
Download | Fast, recommended | |
Download | Best balance | |
Download | Good quality | |
Download | Balanced | |
Download | High quality | |
Download | Very good quality | |
Download | Fast, best quality | |
Download | Maximum accuracy | |
Download | Multimodal projection file for image processing |
π‘ Pro Tip: Start with Q4_K_M
for most use cases, only use F16
if you need maximum precision.
- Downloads last month
- 86
Hardware compatibility
Log In
to view the estimation
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
16-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support
HF Inference deployability: The model has no library tag.
Model tree for matrixportal/X-Ray_Alpha-GGUF
Base model
google/gemma-3-4b-pt
Finetuned
google/gemma-3-4b-it
Finetuned
SicariusSicariiStuff/X-Ray_Alpha