InferenceIllusionist
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -26,7 +26,7 @@ PROUDLY PRESENTS
|
|
26 |
|
27 |
## experiment_1_8b-iMat-GGUF
|
28 |
|
29 |
-
<b>Quantization Note: Use repetition penalty (--repeat-penalty on llama.cpp) of 1.
|
30 |
|
31 |
Quantized from fp16 with love.
|
32 |
* Weighted quantizations were created using fp16 GGUF and [groups_merged-enhancedV2-TurboMini.txt](https://github.com/ggerganov/llama.cpp/discussions/5263#discussioncomment-9432658) in 189 chunks and n_ctx=512
|
|
|
26 |
|
27 |
## experiment_1_8b-iMat-GGUF
|
28 |
|
29 |
+
<b>Quantization Note: Use repetition penalty (--repeat-penalty on llama.cpp) of ~1.15 for best results </b>
|
30 |
|
31 |
Quantized from fp16 with love.
|
32 |
* Weighted quantizations were created using fp16 GGUF and [groups_merged-enhancedV2-TurboMini.txt](https://github.com/ggerganov/llama.cpp/discussions/5263#discussioncomment-9432658) in 189 chunks and n_ctx=512
|