DavidAU commited on
Commit
772ed0a
·
verified ·
1 Parent(s): aef28aa

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +153 -0
README.md ADDED
@@ -0,0 +1,153 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ base_model:
6
+ - Qwen/Qwen2.5-Coder-7B
7
+ - open-r1/OlympicCoder-7B
8
+ pipeline_tag: text-generation
9
+ tags:
10
+ - merge
11
+ - programming
12
+ - code generation
13
+ - code
14
+ - moe
15
+ - mixture of experts
16
+ - qwen2moe
17
+ - 2X11B
18
+ - wolverine coder
19
+ - qwen2
20
+ - codeqwen
21
+ - chat
22
+ - qwen
23
+ - qwen-coder
24
+ - programming
25
+ - code generation
26
+ - code
27
+ - codeqwen
28
+ - moe
29
+ - coding
30
+ - coder
31
+ - qwen2
32
+ - chat
33
+ - qwen
34
+ - qwen-coder
35
+ library_name: transformers
36
+ ---
37
+
38
+ <h2>Qwen2.5-2X11B-CODER-Dueling-Wolverines-V2-28B</h2>
39
+
40
+ <img src="duel-wolverine-coder.jpg" style="float:right; width:300px; height:500px; padding:10px;">
41
+
42
+ "Ripping your programming worries to shreds... fast... times TWO."
43
+
44
+ Tipping the scales at 42 layers and 717 tensors... the monster lives.
45
+
46
+ Two monsters in fact - in one.
47
+
48
+ This is MOE model, using V1 and V2 of Wolverine-Coder 11B which is a merge of two models noted below.
49
+
50
+ This repo has the source code for Version 2.
51
+
52
+ The MOE config gives you full access to both 11B models at full power - full 22B parameters (an additional shared expert brings this to 28B).
53
+
54
+ This MOE model generates stronger, more compact code with an enhanced understanding of your instructions
55
+ and follows what you tell them to the letter.
56
+
57
+ Each 11B version is an overpowered - yet wickedly fast - CODING ENGINE are based on two of the best coder AIs:
58
+
59
+ "Qwen2.5-Coder-7B-Instruct"
60
+
61
+ [ https://huggingface.co/Qwen/Qwen2.5-Coder-7B-Instruct ]
62
+
63
+ and
64
+
65
+ "OlympicCoder-7B"
66
+
67
+ [ https://huggingface.co/open-r1/OlympicCoder-7B ]
68
+
69
+ (11Bs V1 and V2 are here: https://huggingface.co/DavidAU/Qwen2.5-Wolverine-CODER-11B-gguf/ )
70
+
71
+ These two models are stuffed into one compact powerhouse 11BX2 merge that is stronger in performance and understanding
72
+ than both donor models and each of the 11B models created from these too.
73
+
74
+ There are TWO versions of this MOE model too (internal structure is different for each).
75
+
76
+ Quants Q3_K_M and Q6_K are available at the moment, of each version.
77
+
78
+ Limited GGUFS here:
79
+
80
+ https://huggingface.co/DavidAU/Qwen2.5-2X11B-CODER-Dueling-Wolverines-25B-gguf
81
+
82
+ More quants will show under "Quantizations" on the right as they become available.
83
+
84
+ NOTES:
85
+ - Each config/version will be very different from each other.
86
+ - You can select 1 or 2 experts, default is 2 experts.
87
+ - Due to unique setup of this moe, suggest 1-4 generations.
88
+ - Total model size is 28B because Qwen 2.5 MOES have a shared expert in addition to "regular experts"
89
+ - Tool Calling is supported in both versions.
90
+ - Source(s) / full quanting to follow // full repos to follow.
91
+ - Final model size (including layers/tensors) / config subject to change.
92
+
93
+ ---
94
+
95
+ Config / Settings
96
+
97
+ ---
98
+
99
+ Model is set at 32k/32768 context for these GGUFS, full quants/full repos will be 128k/131072.
100
+
101
+ Requirements [Qwen 2.5 7B Coder default settings]:
102
+ - Temp .5 to .7 (or lower)
103
+ - topk: 20, topp: .8, minp: .05
104
+ - rep pen: 1.1 (can be lower)
105
+ - Jinja Template (embedded) or CHATML template.
106
+ - A System Prompt is not required. (ran tests with blank system prompt)
107
+
108
+ Refer to either "Qwen2.5-Coder-7B-Instruct" and/or "OlympicCoder-7B" repos (above) for additional settings, benchmarks and usage.
109
+
110
+ ---
111
+
112
+ <H2>Help, Adjustments, Samplers, Parameters and More</H2>
113
+
114
+ ---
115
+
116
+ <B>Settings: CHAT / ROLEPLAY and/or SMOOTHER operation of this model:</B>
117
+
118
+ In "KoboldCpp" or "oobabooga/text-generation-webui" or "Silly Tavern" ;
119
+
120
+ Set the "Smoothing_factor" to 1.5
121
+
122
+ : in KoboldCpp -> Settings->Samplers->Advanced-> "Smooth_F"
123
+
124
+ : in text-generation-webui -> parameters -> lower right.
125
+
126
+ : In Silly Tavern this is called: "Smoothing"
127
+
128
+
129
+ NOTE: For "text-generation-webui"
130
+
131
+ -> if using GGUFs you need to use "llama_HF" (which involves downloading some config files from the SOURCE version of this model)
132
+
133
+ Source versions (and config files) of my models are here:
134
+
135
+ https://huggingface.co/collections/DavidAU/d-au-source-files-for-gguf-exl2-awq-gptq-hqq-etc-etc-66b55cb8ba25f914cbf210be
136
+
137
+ OTHER OPTIONS:
138
+
139
+ - Increase rep pen to 1.1 to 1.15 (you don't need to do this if you use "smoothing_factor")
140
+
141
+ - If the interface/program you are using to run AI MODELS supports "Quadratic Sampling" ("smoothing") just make the adjustment as noted.
142
+
143
+ <B>Highest Quality Settings / Optimal Operation Guide / Parameters and Samplers</B>
144
+
145
+ This a "Class 1" model:
146
+
147
+ For all settings used for this model (including specifics for its "class"), including example generation(s) and for advanced settings guide (which many times addresses any model issue(s)), including methods to improve model performance for all use case(s) as well as chat, roleplay and other use case(s) please see:
148
+
149
+ [ https://huggingface.co/DavidAU/Maximizing-Model-Performance-All-Quants-Types-And-Full-Precision-by-Samplers_Parameters ]
150
+
151
+ You can see all parameters used for generation, in addition to advanced parameters and samplers to get the most out of this model here:
152
+
153
+ [ https://huggingface.co/DavidAU/Maximizing-Model-Performance-All-Quants-Types-And-Full-Precision-by-Samplers_Parameters ]