SkylarWhite/SahabatAI-MediChatIndo-8B-v1-gguf

This is hosts various GGUF quantized versions of the gmonsoon/SahabatAI-MediChatIndo-8B-v1 model. This model is designed for medical and general-purpose conversational AI in Indonesian and is based on the LLaMA3 architecture. The GGUF format is optimized for efficient inference on low-resource devices and fast deployment.

Model Overview

SahabatAI-MediChatIndo-8B-v1 is a fine-tuned model created by merging:

It has been optimized for understanding and responding in medical and general Indonesian conversations.

GGUF Quantized Versions

The following GGUF quantized versions are available in this repository:

  • 16-bit (F16): High-precision quantization for use cases requiring maximal accuracy.
  • Q4_K_M: Balanced between speed and performance, ideal for most use cases.
  • Q5_K_M: Improved precision over Q4 while maintaining efficient performance.
  • Q8_0: Full precision for demanding tasks where accuracy is critical.

Feedback and Contributions

Feedback and contributions are welcome! Please open an issue or contact the model's author for further discussions.

Downloads last month
13
GGUF
Model size
8.03B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

4-bit

5-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for SkylarWhite/SahabatAI-MediChatIndo-8B-v1-gguf

Quantized
(3)
this model

Collection including SkylarWhite/SahabatAI-MediChatIndo-8B-v1-gguf