|
--- |
|
library_name: transformers |
|
language: |
|
- en |
|
- fr |
|
- it |
|
- pt |
|
- hi |
|
- es |
|
- th |
|
- de |
|
base_model: |
|
- ockerman0/MN-12B-Starcannon-v5.5-unofficial |
|
base_model_relation: quantized |
|
tags: |
|
- mergekit |
|
- merge |
|
- mistral |
|
quantized_by: tsss |
|
pipeline_tag: text-generation |
|
--- |
|
|
|
## This repo contains EXL2 quants of ockerman0/MN-12B-Starcannon-v5.5-unofficial. |
|
|
|
Find the original model card [here](https://huggingface.co/ockerman0/MN-12B-Starcannon-v5.5-unofficial). |
|
|
|
## Base repo only contains the measurement file, see revisions for the quants. |
|
|
|
- [measurement.json](https://huggingface.co/tssst/MN-12B-Starcannon-v5.5-unofficial-EXL2/tree/main) |
|
- [3.0bpw](https://huggingface.co/tssst/MN-12B-Starcannon-v5.5-unofficial-EXL2/tree/3bpw) |
|
- [4.0bpw](https://huggingface.co/tssst/MN-12B-Starcannon-v5.5-unofficial-EXL2/tree/4bpw) |
|
- [5.0bpw](https://huggingface.co/tssst/MN-12B-Starcannon-v5.5-unofficial-EXL2/tree/5bpw) |
|
- [6.0bpw](https://huggingface.co/tssst/MN-12B-Starcannon-v5.5-unofficial-EXL2/tree/6bpw) |
|
|
|
## Notes |
|
Making these was a lesson in pain and humility. It has been over two months since the day I decided "hm today i will learn how to make exl2 quants" <- (clueless). First my conda env |
|
stopped working (for some reason), then it stopped recognizing venvs when I tried using those, then the universe decided to screw up the one venv I had working somehow (I can only assume |
|
it was a cosmic bitflip or something because it literally stopped working overnight) and making these four quants alone took over an hour on my hardware, in which time I could probably have |
|
made an entire set of GGUFs (plus a full set of i-quants) for three different models. Then uploading these was such a pain because huggingface-cli might as well be arcane magic since their |
|
documentation doesn't really tell you how to actually use it or what exactly will happen when you run this method. I haven't even tested any of these quants because tabbyapi, to this day, |
|
simply will not work. Torch keeps bugging me about running out of VRAM, even when trying to load 3bs. I have basically tried everything to try and get tabbyapi to run. It simply will not. |
|
|
|
Suggest more models in the community tab and I might have a crack at exl2'ing them. |
|
|
|
The model is quite nice though, it is quite useful for my usecases of synthetic variant generation if you're into that sort of thing. |