Text Generation
GGUF
English
reasoning
thinking
uncensored
gated
mixture of experts
Mixture of Experts
8x3B
Llama 3.2 MOE
128k context
creative
creative writing
fiction writing
plot generation
sub-plot generation
story generation
scene continue
storytelling
fiction story
science fiction
romance
all genres
story
writing
vivid prosing
vivid writing
fiction
roleplaying
float32
swearing
rp
horror
mergekit
Update README.md
Browse files
README.md
CHANGED
@@ -62,7 +62,8 @@ and quants have the output tensor at Q8_0, with a few choice quants at f16 (16 b
|
|
62 |
The "gated" strucure means the "reasoning model" is re-inforced by the other 7 models in the MOE during reasoning, and then during
|
63 |
output generation / non-reasoning the non-reasoning model(s) take control.
|
64 |
|
65 |
-
Also, with "gating" you can directly access/control the model(s) you want to use during instruction following and generation.
|
|
|
66 |
|
67 |
This model is the "reasoning / gated version" of this model:
|
68 |
|
|
|
62 |
The "gated" strucure means the "reasoning model" is re-inforced by the other 7 models in the MOE during reasoning, and then during
|
63 |
output generation / non-reasoning the non-reasoning model(s) take control.
|
64 |
|
65 |
+
Also, with "gating" you can directly access/control the model(s) you want to use during instruction following and generation. This
|
66 |
+
is optional - you don't need to set this to use the model normally.
|
67 |
|
68 |
This model is the "reasoning / gated version" of this model:
|
69 |
|