Text Generation
GGUF
English
Chinese
Cubed Reasoning
QwQ-32B
reasoning
thinking
r1
cot
deepseek
Qwen2.5
Hermes
DeepHermes
DeepSeek
DeepSeek-R1-Distill
Uncensored
creative
128k context
general usage
problem solving
brainstorming
solve riddles
fiction writing
plot generation
sub-plot generation
story generation
scene continue
storytelling
fiction story
story
writing
fiction
roleplaying
swearing
horror
Qwen 2.5
mergekit
conversational
Update README.md
Browse files
README.md
CHANGED
@@ -151,6 +151,20 @@ Record so far (mine): 12k output (coherent) with 4k context limit.
|
|
151 |
For some AI apps use of the Jinja Template (embedded in the GGUFs) may not work, and you need to manual select/use "ChatML" template
|
152 |
in your AI/LLM app.
|
153 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
154 |
<b>Optional : Rocket Fuel for Thought</b>
|
155 |
|
156 |
This is an optional system prompt you can use to enhance both "thinking/reasoning" and "output".
|
|
|
151 |
For some AI apps use of the Jinja Template (embedded in the GGUFs) may not work, and you need to manual select/use "ChatML" template
|
152 |
in your AI/LLM app.
|
153 |
|
154 |
+
<B>Quant Choice Notes:</b>
|
155 |
+
|
156 |
+
This model shows much stronger detail, generation and thoughts/reasoning as you go up in quant(s).
|
157 |
+
|
158 |
+
In terms of "reasoning/thinking" length this can be HALVED for some "problems" even if you go from Q2k up to Q3KM.
|
159 |
+
|
160 |
+
I.E: It figures out the solution to the problem faster.
|
161 |
+
|
162 |
+
Likewise, detail in output as well as detail in reasoning will be deeper and stronger.
|
163 |
+
|
164 |
+
With that in mind, even Q2k (the smallest/lowest regular quant) is potent.
|
165 |
+
|
166 |
+
Also, the same quant in "Imatrix" maybe even stronger than the regular version.
|
167 |
+
|
168 |
<b>Optional : Rocket Fuel for Thought</b>
|
169 |
|
170 |
This is an optional system prompt you can use to enhance both "thinking/reasoning" and "output".
|