gguf when? cmon, its been 11 min already!
lol well darn, i had plans today... oof... as a quantizer, i wonder if i should wait for the -Instruct
? is that out yet? lol...
+1 gguf please
wait for instruct model, not sure how gguf of the base model could be usefull for personal usage
Base models are good for creative writing.
lol well darn, i had plans today... oof... as a quantizer, i wonder if i should wait for the
-Instruct
? is that out yet? lol...
How dare you have plans when ds puts out a new model!!! 😂
"Why is the GGUF so late it's been 20 seconds already!"
i think lets wait for instruct version. I am very patient. very very very patient.
I think llama.cpp needs to be updated first.
I think llama.cpp needs to be updated first.
https://huggingface.co/deepseek-ai/DeepSeek-V3-Base/blob/main/model.safetensors.index.json
https://huggingface.co/deepseek-ai/DeepSeek-V3.1-Base/blob/main/model.safetensors.index.json
These look identical
I figured out how to create the bf16 safetensors, now I'm creating the bf16 gguf. We'll see.
Yeah, seems like it needs some changes to llama.cpp. I got it inferring but the chat template seems messed up.
I'm throwing a Q4_K_M up soon while I work on imatrix and further quants
@createthis it's also a base model so chatting is not going to be as reliable without giving it a multi turn prompt
https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3.1-Base-Q4_K_M-GGUF
In case anyone wants to try Q4_K_M