InferenceIllusionist
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -6,12 +6,15 @@ tags:
|
|
6 |
- conversational
|
7 |
- gguf
|
8 |
---
|
|
|
9 |
|
10 |
[CybersurferNyandroidLexicat](https://huggingface.co/Envoid/CybersurferNyandroidLexicat-8x7B) quantized from fp16 with love.
|
11 |
|
12 |
Uses the same imat calculation method as the later batch of maid-yuzu-v8-alter-iMat-GGUF.
|
13 |
|
14 |
-
<b>Legacy quants (i.e. Q5_K_M, Q6_K, etc) in this repo have all been enhanced with the imatrix calculation.
|
|
|
|
|
15 |
|
16 |
For more information on latest iMatrix quants see this PR - https://github.com/ggerganov/llama.cpp/pull/5747
|
17 |
|
|
|
6 |
- conversational
|
7 |
- gguf
|
8 |
---
|
9 |
+
# CybersurferNyandroidLexicat-8x7B-iMat-GGUF
|
10 |
|
11 |
[CybersurferNyandroidLexicat](https://huggingface.co/Envoid/CybersurferNyandroidLexicat-8x7B) quantized from fp16 with love.
|
12 |
|
13 |
Uses the same imat calculation method as the later batch of maid-yuzu-v8-alter-iMat-GGUF.
|
14 |
|
15 |
+
<b>Legacy quants (i.e. Q5_K_M, Q6_K, etc) in this repo have all been enhanced with the imatrix calculation. These quants show improved KL-Divergence over their static counterparts.</b>
|
16 |
+
|
17 |
+
All files included here for your convenience. No need to clone the entire repo, just pick the quant that's right for you.
|
18 |
|
19 |
For more information on latest iMatrix quants see this PR - https://github.com/ggerganov/llama.cpp/pull/5747
|
20 |
|