X-Ray_Alpha-GGUF / README.md
matrixportal's picture
Update README.md
a1fd207 verified
metadata
base_model: SicariusSicariiStuff/X-Ray_Alpha
language:
  - en
license: gemma
tags:
  - llama-cpp
  - matrixportal

matrixportal/X-Ray_Alpha-GGUF

This model was converted to GGUF format from SicariusSicariiStuff/X-Ray_Alpha using llama.cpp via the ggml.ai's all-gguf-same-where space. Refer to the original model card for more details on the model.

βœ… Quantized Models Download List

πŸ” Recommended Quantizations

  • ✨ General CPU Use: Q4_K_M (Best balance of speed/quality)
  • πŸ“± ARM Devices: Q4_0 (Optimized for ARM CPUs)
  • πŸ† Maximum Quality: Q8_0 (Near-original quality)

πŸ“¦ Full Quantization Options

πŸš€ Download πŸ”’ Type πŸ“ Notes
Download Q2_K Basic quantization
Download Q3_K_S Small size
Download Q3_K_M Balanced quality
Download Q3_K_L Better quality
Download Q4_0 Fast on ARM
Download Q4_K_S Fast, recommended
Download Q4_K_M ⭐ Best balance
Download Q5_0 Good quality
Download Q5_K_S Balanced
Download Q5_K_M High quality
Download Q6_K πŸ† Very good quality
Download Q8_0 ⚑ Fast, best quality
Download F16 Maximum accuracy
Download mmproj Multimodal projection file for image processing

πŸ’‘ Pro Tip: Start with Q4_K_M for most use cases, only use F16 if you need maximum precision.