File size: 2,347 Bytes
165fdb5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
94cbf20
165fdb5
 
 
 
 
 
 
 
 
94cbf20
7c7c45c
 
 
 
165fdb5
 
 
 
 
 
 
 
 
 
 
42a3c6e
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
---
library_name: transformers
language:
- en
- fr
- it
- pt
- hi
- es
- th
- de
base_model:
- ockerman0/MN-12B-Starcannon-v5.5-unofficial
base_model_relation: quantized
tags:
- mergekit
- merge
- mistral
quantized_by: tsss
pipeline_tag: text-generation
---

## This repo contains EXL2 quants of ockerman0/MN-12B-Starcannon-v5.5-unofficial. 

Find the original model card [here](https://huggingface.co/ockerman0/MN-12B-Starcannon-v5.5-unofficial).

## Base repo only contains the measurement file, see revisions for the quants.

- [measurement.json](https://huggingface.co/tssst/MN-12B-Starcannon-v5.5-unofficial-EXL2/tree/main)
- [3.0bpw](https://huggingface.co/tssst/MN-12B-Starcannon-v5.5-unofficial-EXL2/tree/3bpw)
- [4.0bpw](https://huggingface.co/tssst/MN-12B-Starcannon-v5.5-unofficial-EXL2/tree/4bpw)
- [5.0bpw](https://huggingface.co/tssst/MN-12B-Starcannon-v5.5-unofficial-EXL2/tree/5bpw)
- [6.0bpw](https://huggingface.co/tssst/MN-12B-Starcannon-v5.5-unofficial-EXL2/tree/6bpw)

## Notes
Making these was a lesson in pain and humility. It has been over two months since the day I decided "hm today i will learn how to make exl2 quants" <- (clueless). First my conda env 
stopped working (for some reason), then it stopped recognizing venvs when I tried using those, then the universe decided to screw up the one venv I had working somehow (I can only assume 
it was a cosmic bitflip or something because it literally stopped working overnight) and making these four quants alone took over an hour on my hardware, in which time I could probably have 
made an entire set of GGUFs (plus a full set of i-quants) for three different models. Then uploading these was such a pain because huggingface-cli might as well be arcane magic since their 
documentation doesn't really tell you how to actually use it or what exactly will happen when you run this method. I haven't even tested any of these quants because tabbyapi, to this day, 
simply will not work. Torch keeps bugging me about running out of VRAM, even when trying to load 3bs. I have basically tried everything to try and get tabbyapi to run. It simply will not.

Suggest more models in the community tab and I might have a crack at exl2'ing them.

The model is quite nice though, it is quite useful for my usecases of synthetic variant generation if you're into that sort of thing.