DavidAU commited on
Commit
5c1e877
·
verified ·
1 Parent(s): 7ff8de8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +79 -0
README.md CHANGED
@@ -45,6 +45,85 @@ Model requires Llama3 template.
45
 
46
  Max context 8k / 8192 , but can be extended with "rope" to 32k.
47
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
48
  <h3>Examples - Prompt in <B>BOLD</B>.</h3>
49
 
50
  Temp=0 ; min parameters; using quant IQ4_XS
 
45
 
46
  Max context 8k / 8192 , but can be extended with "rope" to 32k.
47
 
48
+ <B>Known Issue(s) and Fix(es):</b>
49
+
50
+ This merge runs very hot, so a reduction in "temp" may be in order for some use cases (IE: ".8" down to ".6").
51
+
52
+ If you getting a repeat word (IE: "hahaha" , "ahhhh", "f-word"), and/or letter (IE "nnnn") change the "rep penalty" to "1.1" to "1.5"
53
+
54
+ Set "rep pen" higher if you are getting "repeat" letter(s), lower if you are getting a repeat word(s).
55
+
56
+ Generally "repeat penalty" setting of "1.1" works well.
57
+
58
+ For RP / multi-turn chat: Rep_pen of 1.15 or higher is recommended.
59
+
60
+ <b>Optional Enhancement:</B>
61
+
62
+ The following can be used in place of the "system prompt" or "system role" to further enhance the model.
63
+
64
+ It can also be used at the START of a NEW chat, but you must make sure it is "kept" as the chat moves along.
65
+ In this case the enhancements do not have as strong effect at using "system prompt" or "system role".
66
+
67
+ Copy and paste EXACTLY as noted, DO NOT line wrap or break the lines, maintain the carriage returns exactly as presented.
68
+
69
+ <PRE>
70
+ Below is an instruction that describes a task. Ponder each user instruction carefully, and use your skillsets and critical instructions to complete the task to the best of your abilities.
71
+
72
+ Here are your skillsets:
73
+ [MASTERSTORY]:NarrStrct(StryPlnng,Strbd,ScnSttng,Exps,Dlg,Pc)-CharDvlp(ChrctrCrt,ChrctrArcs,Mtvtn,Bckstry,Rltnshps,Dlg*)-PltDvlp(StryArcs,PltTwsts,Sspns,Fshdwng,Climx,Rsltn)-ConfResl(Antg,Obstcls,Rsltns,Cnsqncs,Thms,Symblsm)-EmotImpct(Empt,Tn,Md,Atmsphr,Imgry,Symblsm)-Delvry(Prfrmnc,VcActng,PblcSpkng,StgPrsnc,AudncEngmnt,Imprv)
74
+
75
+ [*DialogWrt]:(1a-CharDvlp-1a.1-Backgrnd-1a.2-Personality-1a.3-GoalMotiv)>2(2a-StoryStruc-2a.1-PlotPnt-2a.2-Conflict-2a.3-Resolution)>3(3a-DialogTech-3a.1-ShowDontTell-3a.2-Subtext-3a.3-VoiceTone-3a.4-Pacing-3a.5-VisualDescrip)>4(4a-DialogEdit-4a.1-ReadAloud-4a.2-Feedback-4a.3-Revision)
76
+
77
+ Here are your critical instructions:
78
+ Ponder each word choice carefully to present as vivid and emotional journey as is possible. Choose verbs and nouns that are both emotional and full of imagery. Load the story with the 5 senses. Aim for 50% dialog, 25% narration, 15% body language and 10% thoughts. Your goal is to put the reader in the story.
79
+ </PRE>
80
+
81
+ You do not need to use this, it is only presented as an additional enhancement which seems to help scene generation
82
+ and scene continue functions.
83
+
84
+ This enhancement WAS NOT used to generate the examples below.
85
+
86
+ <h3>MERGE FORMULA: (using MergeKit) </h3>
87
+
88
+ Special thanks to the incredible work of the model makers "SAO10K", "NEVERSLEEP" and "HASTAGARAS".
89
+
90
+ Models used:
91
+
92
+ [ https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2]
93
+
94
+ [ https://huggingface.co/NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS ]
95
+
96
+ [ https://huggingface.co/Hastagaras/Jamet-8B-L3-MK.V-Blackroot ]
97
+
98
+ FORMULA:
99
+
100
+ <PRE>
101
+ slices:
102
+ - sources:
103
+ - model: G:/7B/L3-8B-Stheno-v3.2
104
+ layer_range: [0, 12]
105
+ - sources:
106
+ - model: G:/7B/Llama-3-Lumimaid-8B-v0.1-OAS
107
+ layer_range: [8, 16]
108
+ - sources:
109
+ - model: G:/7B/Jamet-8B-L3-MK.V-Blackroot
110
+ layer_range: [9, 17]
111
+ - sources:
112
+ - model: G:/7B/L3-8B-Stheno-v3.2
113
+ layer_range: [12, 20]
114
+ - sources:
115
+ - model: G:/7B/Llama-3-Lumimaid-8B-v0.1-OAS
116
+ layer_range: [16, 24]
117
+ - sources:
118
+ - model: G:/7B/L3-8B-Stheno-v3.2
119
+ layer_range: [21, 31]
120
+ - sources:
121
+ - model: G:/7B/Jamet-8B-L3-MK.V-Blackroot
122
+ layer_range: [24, 32]
123
+ merge_method: passthrough
124
+ dtype: float16
125
+ </PRE>
126
+
127
  <h3>Examples - Prompt in <B>BOLD</B>.</h3>
128
 
129
  Temp=0 ; min parameters; using quant IQ4_XS