QuantFactory/Lumimaid-Magnum-v4-12B-GGUF
This is quantized version of Undi95/Lumimaid-Magnum-v4-12B created using llama.cpp
Original Model Card
Merge of Lumimaid and Magnum as requested by some. UPDATE : Magnum v4 used in this merge as asked here
I used the DELLA merge method in mergekit and added a finetune of Nemo only on Claude input, trained on 16k ctx, in the mix.
Prompt template: Mistral
<s>[INST] {input} [/INST] {output}</s>
- Downloads last month
- 350
Hardware compatibility
Log In
to view the estimation
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
HF Inference deployability: The model has no pipeline_tag.