InferenceIllusionist
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -7,7 +7,7 @@ license: apache-2.0
|
|
7 |
# Hyperion-1.5-Mistral-7B-iMat-GGUF
|
8 |
|
9 |
New importance matrix quantizations for Hyperion-1.5-Mistral-7B.
|
10 |
-
These i-quants have a better size to perplexity ratio as they were creating using an Importance Matrix file
|
11 |
|
12 |
<b>All files created using latest (3/2) llama.cpp build, including IQ3_S improvements covered [here](https://github.com/ggerganov/llama.cpp/pull/5829)</b>
|
13 |
|
|
|
7 |
# Hyperion-1.5-Mistral-7B-iMat-GGUF
|
8 |
|
9 |
New importance matrix quantizations for Hyperion-1.5-Mistral-7B.
|
10 |
+
These i-quants have a better size to perplexity ratio as they were creating using an Importance Matrix file calculated from the fp16 (unquantized) gguf.
|
11 |
|
12 |
<b>All files created using latest (3/2) llama.cpp build, including IQ3_S improvements covered [here](https://github.com/ggerganov/llama.cpp/pull/5829)</b>
|
13 |
|