Custom GGUF quants of NousResearch/DeepHermes-3-Mistral-24B-Preview, where the Output Tensors are kept at F32 or quantized to Q8_0, and the Embeddings are kept at F32 or Quantized to Q8_0. Enjoy! 🧠🔥🚀
Custom GGUF quants of NousResearch/DeepHermes-3-Mistral-24B-Preview, where the Output Tensors are kept at F32 or quantized to Q8_0, and the Embeddings are kept at F32 or Quantized to Q8_0. Enjoy! 🧠🔥🚀