DavidAU commited on
Commit
51dd318
·
verified ·
1 Parent(s): 968d199

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +27 -0
README.md CHANGED
@@ -161,6 +161,33 @@ Regular and Special Usage:
161
  You may want to use "regular" Dark Planet 8B [https://huggingface.co/DavidAU/L3-Dark-Planet-8B-GGUF] for some writing task(s),
162
  and this model for prose specific task(s).
163
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
164
  <B>Model Template:</B>
165
 
166
  This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 8k / 8192.
 
161
  You may want to use "regular" Dark Planet 8B [https://huggingface.co/DavidAU/L3-Dark-Planet-8B-GGUF] for some writing task(s),
162
  and this model for prose specific task(s).
163
 
164
+ <B>Settings: CHAT / ROLEPLAY and/or SMOOTHER operation of this model:</B>
165
+
166
+ In "KoboldCpp" or "oobabooga/text-generation-webui" or "Silly Tavern" ;
167
+
168
+ Set the "Smoothing_factor" to 1.5 to 2.5
169
+
170
+ : in KoboldCpp -> Settings->Samplers->Advanced-> "Smooth_F"
171
+
172
+ : in text-generation-webui -> parameters -> lower right.
173
+
174
+ : In Silly Tavern this is called: "Smoothing"
175
+
176
+
177
+ NOTE: For "text-generation-webui"
178
+
179
+ -> if using GGUFs you need to use "llama_HF" (which involves downloading some config files from the SOURCE version of this model)
180
+
181
+ Source versions (and config files) of my models are here:
182
+
183
+ https://huggingface.co/collections/DavidAU/d-au-source-files-for-gguf-exl2-awq-gptq-hqq-etc-etc-66b55cb8ba25f914cbf210be
184
+
185
+ OTHER OPTIONS:
186
+
187
+ - Increase rep pen to 1.1 to 1.15 (you don't need to do this if you use "smoothing_factor")
188
+
189
+ - If the interface/program you are using to run AI MODELS supports "Quadratic Sampling" ("smoothing") just make the adjustment as noted.
190
+
191
  <B>Model Template:</B>
192
 
193
  This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 8k / 8192.