GGUF
conversational

Sarvam-M

Chat on Sarvam Playground

Model Information

This repository contains gguf version of sarvam-m in q8 precision.

Learn more about sarvam-m in our detailed blog post.

Running the model on a CPU

You can use the model on your local machine (without gpu) as explained here.

Example Command:

./build/bin/llama-cli -i -m /your/folder/path/sarvam-m-q8_0.gguf -c 8192 -t 16
Downloads last month
171
GGUF
Model size
23.6B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for sarvamai/sarvam-m-q8-gguf

Finetuned
sarvamai/sarvam-m
Quantized
(17)
this model

Collection including sarvamai/sarvam-m-q8-gguf