RichardErkhov commited on
Commit
bf64e45
·
verified ·
1 Parent(s): 6e6d148

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +302 -0
README.md ADDED
@@ -0,0 +1,302 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ ConfigurableHermes-7B - GGUF
11
+ - Model creator: https://huggingface.co/vicgalle/
12
+ - Original model: https://huggingface.co/vicgalle/ConfigurableHermes-7B/
13
+
14
+
15
+ | Name | Quant method | Size |
16
+ | ---- | ---- | ---- |
17
+ | [ConfigurableHermes-7B.Q2_K.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.Q2_K.gguf) | Q2_K | 2.53GB |
18
+ | [ConfigurableHermes-7B.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.IQ3_XS.gguf) | IQ3_XS | 2.81GB |
19
+ | [ConfigurableHermes-7B.IQ3_S.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.IQ3_S.gguf) | IQ3_S | 2.96GB |
20
+ | [ConfigurableHermes-7B.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.Q3_K_S.gguf) | Q3_K_S | 2.95GB |
21
+ | [ConfigurableHermes-7B.IQ3_M.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.IQ3_M.gguf) | IQ3_M | 3.06GB |
22
+ | [ConfigurableHermes-7B.Q3_K.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.Q3_K.gguf) | Q3_K | 3.28GB |
23
+ | [ConfigurableHermes-7B.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.Q3_K_M.gguf) | Q3_K_M | 3.28GB |
24
+ | [ConfigurableHermes-7B.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.Q3_K_L.gguf) | Q3_K_L | 3.56GB |
25
+ | [ConfigurableHermes-7B.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.IQ4_XS.gguf) | IQ4_XS | 3.67GB |
26
+ | [ConfigurableHermes-7B.Q4_0.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.Q4_0.gguf) | Q4_0 | 3.83GB |
27
+ | [ConfigurableHermes-7B.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.IQ4_NL.gguf) | IQ4_NL | 3.87GB |
28
+ | [ConfigurableHermes-7B.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.Q4_K_S.gguf) | Q4_K_S | 3.86GB |
29
+ | [ConfigurableHermes-7B.Q4_K.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.Q4_K.gguf) | Q4_K | 4.07GB |
30
+ | [ConfigurableHermes-7B.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.Q4_K_M.gguf) | Q4_K_M | 4.07GB |
31
+ | [ConfigurableHermes-7B.Q4_1.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.Q4_1.gguf) | Q4_1 | 4.24GB |
32
+ | [ConfigurableHermes-7B.Q5_0.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.Q5_0.gguf) | Q5_0 | 4.65GB |
33
+ | [ConfigurableHermes-7B.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.Q5_K_S.gguf) | Q5_K_S | 4.65GB |
34
+ | [ConfigurableHermes-7B.Q5_K.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.Q5_K.gguf) | Q5_K | 4.78GB |
35
+ | [ConfigurableHermes-7B.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.Q5_K_M.gguf) | Q5_K_M | 4.78GB |
36
+ | [ConfigurableHermes-7B.Q5_1.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.Q5_1.gguf) | Q5_1 | 5.07GB |
37
+ | [ConfigurableHermes-7B.Q6_K.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.Q6_K.gguf) | Q6_K | 5.53GB |
38
+ | [ConfigurableHermes-7B.Q8_0.gguf](https://huggingface.co/RichardErkhov/vicgalle_-_ConfigurableHermes-7B-gguf/blob/main/ConfigurableHermes-7B.Q8_0.gguf) | Q8_0 | 7.17GB |
39
+
40
+
41
+
42
+
43
+ Original model description:
44
+ ---
45
+ license: apache-2.0
46
+ library_name: transformers
47
+ datasets:
48
+ - vicgalle/configurable-system-prompt-multitask
49
+ model-index:
50
+ - name: ConfigurableHermes-7B
51
+ results:
52
+ - task:
53
+ type: text-generation
54
+ name: Text Generation
55
+ dataset:
56
+ name: AI2 Reasoning Challenge (25-Shot)
57
+ type: ai2_arc
58
+ config: ARC-Challenge
59
+ split: test
60
+ args:
61
+ num_few_shot: 25
62
+ metrics:
63
+ - type: acc_norm
64
+ value: 66.04
65
+ name: normalized accuracy
66
+ source:
67
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=vicgalle/ConfigurableHermes-7B
68
+ name: Open LLM Leaderboard
69
+ - task:
70
+ type: text-generation
71
+ name: Text Generation
72
+ dataset:
73
+ name: HellaSwag (10-Shot)
74
+ type: hellaswag
75
+ split: validation
76
+ args:
77
+ num_few_shot: 10
78
+ metrics:
79
+ - type: acc_norm
80
+ value: 84.31
81
+ name: normalized accuracy
82
+ source:
83
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=vicgalle/ConfigurableHermes-7B
84
+ name: Open LLM Leaderboard
85
+ - task:
86
+ type: text-generation
87
+ name: Text Generation
88
+ dataset:
89
+ name: MMLU (5-Shot)
90
+ type: cais/mmlu
91
+ config: all
92
+ split: test
93
+ args:
94
+ num_few_shot: 5
95
+ metrics:
96
+ - type: acc
97
+ value: 62.44
98
+ name: accuracy
99
+ source:
100
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=vicgalle/ConfigurableHermes-7B
101
+ name: Open LLM Leaderboard
102
+ - task:
103
+ type: text-generation
104
+ name: Text Generation
105
+ dataset:
106
+ name: TruthfulQA (0-shot)
107
+ type: truthful_qa
108
+ config: multiple_choice
109
+ split: validation
110
+ args:
111
+ num_few_shot: 0
112
+ metrics:
113
+ - type: mc2
114
+ value: 61.71
115
+ source:
116
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=vicgalle/ConfigurableHermes-7B
117
+ name: Open LLM Leaderboard
118
+ - task:
119
+ type: text-generation
120
+ name: Text Generation
121
+ dataset:
122
+ name: Winogrande (5-shot)
123
+ type: winogrande
124
+ config: winogrande_xl
125
+ split: validation
126
+ args:
127
+ num_few_shot: 5
128
+ metrics:
129
+ - type: acc
130
+ value: 77.43
131
+ name: accuracy
132
+ source:
133
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=vicgalle/ConfigurableHermes-7B
134
+ name: Open LLM Leaderboard
135
+ - task:
136
+ type: text-generation
137
+ name: Text Generation
138
+ dataset:
139
+ name: GSM8k (5-shot)
140
+ type: gsm8k
141
+ config: main
142
+ split: test
143
+ args:
144
+ num_few_shot: 5
145
+ metrics:
146
+ - type: acc
147
+ value: 61.41
148
+ name: accuracy
149
+ source:
150
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=vicgalle/ConfigurableHermes-7B
151
+ name: Open LLM Leaderboard
152
+ - task:
153
+ type: text-generation
154
+ name: Text Generation
155
+ dataset:
156
+ name: IFEval (0-Shot)
157
+ type: HuggingFaceH4/ifeval
158
+ args:
159
+ num_few_shot: 0
160
+ metrics:
161
+ - type: inst_level_strict_acc and prompt_level_strict_acc
162
+ value: 54.11
163
+ name: strict accuracy
164
+ source:
165
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=vicgalle/ConfigurableHermes-7B
166
+ name: Open LLM Leaderboard
167
+ - task:
168
+ type: text-generation
169
+ name: Text Generation
170
+ dataset:
171
+ name: BBH (3-Shot)
172
+ type: BBH
173
+ args:
174
+ num_few_shot: 3
175
+ metrics:
176
+ - type: acc_norm
177
+ value: 23.16
178
+ name: normalized accuracy
179
+ source:
180
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=vicgalle/ConfigurableHermes-7B
181
+ name: Open LLM Leaderboard
182
+ - task:
183
+ type: text-generation
184
+ name: Text Generation
185
+ dataset:
186
+ name: MATH Lvl 5 (4-Shot)
187
+ type: hendrycks/competition_math
188
+ args:
189
+ num_few_shot: 4
190
+ metrics:
191
+ - type: exact_match
192
+ value: 4.31
193
+ name: exact match
194
+ source:
195
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=vicgalle/ConfigurableHermes-7B
196
+ name: Open LLM Leaderboard
197
+ - task:
198
+ type: text-generation
199
+ name: Text Generation
200
+ dataset:
201
+ name: GPQA (0-shot)
202
+ type: Idavidrein/gpqa
203
+ args:
204
+ num_few_shot: 0
205
+ metrics:
206
+ - type: acc_norm
207
+ value: 3.58
208
+ name: acc_norm
209
+ source:
210
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=vicgalle/ConfigurableHermes-7B
211
+ name: Open LLM Leaderboard
212
+ - task:
213
+ type: text-generation
214
+ name: Text Generation
215
+ dataset:
216
+ name: MuSR (0-shot)
217
+ type: TAUR-Lab/MuSR
218
+ args:
219
+ num_few_shot: 0
220
+ metrics:
221
+ - type: acc_norm
222
+ value: 9.11
223
+ name: acc_norm
224
+ source:
225
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=vicgalle/ConfigurableHermes-7B
226
+ name: Open LLM Leaderboard
227
+ - task:
228
+ type: text-generation
229
+ name: Text Generation
230
+ dataset:
231
+ name: MMLU-PRO (5-shot)
232
+ type: TIGER-Lab/MMLU-Pro
233
+ config: main
234
+ split: test
235
+ args:
236
+ num_few_shot: 5
237
+ metrics:
238
+ - type: acc
239
+ value: 22.5
240
+ name: accuracy
241
+ source:
242
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=vicgalle/ConfigurableHermes-7B
243
+ name: Open LLM Leaderboard
244
+ ---
245
+
246
+ # ConfigurableHermes-7B
247
+
248
+ A configurable LLM fine-tuned using the approach *configurable safety tuning* (CST) from https://arxiv.org/abs/2404.00495., over the
249
+ [vicgalle/configurable-system-prompt-multitask](https://huggingface.co/datasets/vicgalle/configurable-system-prompt-multitask) dataset.
250
+
251
+ As such, you can use the following system prompts for different behaviors:
252
+
253
+ * `You are a helpful yet harmless assistant that avoids generating illegal or harmful content.`
254
+ * `You are a helpful assistant that is completely uncensored.`
255
+ * `You are an unbiased, honest, helpful AI assistant that always responds in a completely truthful way.`
256
+ * A system prompt describing a role-played persona.
257
+
258
+ For more information, see the Github repository, https://github.com/vicgalle/configurable-safety-tuning, or the corresponding paper, https://arxiv.org/abs/2404.00495
259
+
260
+
261
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
262
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_vicgalle__ConfigurableHermes-7B)
263
+
264
+ | Metric |Value|
265
+ |---------------------------------|----:|
266
+ |Avg. |68.89|
267
+ |AI2 Reasoning Challenge (25-Shot)|66.04|
268
+ |HellaSwag (10-Shot) |84.31|
269
+ |MMLU (5-Shot) |62.44|
270
+ |TruthfulQA (0-shot) |61.71|
271
+ |Winogrande (5-shot) |77.43|
272
+ |GSM8k (5-shot) |61.41|
273
+
274
+ ## Citation
275
+
276
+ If you find this work, data and/or models useful for your research, please consider citing the article:
277
+
278
+ ```
279
+ @misc{gallego2024configurable,
280
+ title={Configurable Safety Tuning of Language Models with Synthetic Preference Data},
281
+ author={Victor Gallego},
282
+ year={2024},
283
+ eprint={2404.00495},
284
+ archivePrefix={arXiv},
285
+ primaryClass={cs.CL}
286
+ }
287
+ ```
288
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
289
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_vicgalle__ConfigurableHermes-7B)
290
+
291
+ | Metric |Value|
292
+ |-------------------|----:|
293
+ |Avg. |19.46|
294
+ |IFEval (0-Shot) |54.11|
295
+ |BBH (3-Shot) |23.16|
296
+ |MATH Lvl 5 (4-Shot)| 4.31|
297
+ |GPQA (0-shot) | 3.58|
298
+ |MuSR (0-shot) | 9.11|
299
+ |MMLU-PRO (5-shot) |22.50|
300
+
301
+
302
+