itlwas commited on
Commit
283b144
·
verified ·
1 Parent(s): 14e8e17

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +292 -0
README.md ADDED
@@ -0,0 +1,292 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: other
5
+ tags:
6
+ - axolotl
7
+ - generated_from_trainer
8
+ - Mistral
9
+ - instruct
10
+ - finetune
11
+ - chatml
12
+ - gpt4
13
+ - synthetic data
14
+ - science
15
+ - physics
16
+ - chemistry
17
+ - biology
18
+ - math
19
+ - llama-cpp
20
+ - gguf-my-repo
21
+ base_model: Weyaxi/Einstein-v4-7B
22
+ datasets:
23
+ - allenai/ai2_arc
24
+ - camel-ai/physics
25
+ - camel-ai/chemistry
26
+ - camel-ai/biology
27
+ - camel-ai/math
28
+ - metaeval/reclor
29
+ - openbookqa
30
+ - mandyyyyii/scibench
31
+ - derek-thomas/ScienceQA
32
+ - TIGER-Lab/ScienceEval
33
+ - jondurbin/airoboros-3.2
34
+ - LDJnr/Capybara
35
+ - Cot-Alpaca-GPT4-From-OpenHermes-2.5
36
+ - STEM-AI-mtl/Electrical-engineering
37
+ - knowrohit07/saraswati-stem
38
+ - sablo/oasst2_curated
39
+ - glaiveai/glaive-code-assistant
40
+ - lmsys/lmsys-chat-1m
41
+ - TIGER-Lab/MathInstruct
42
+ - bigbio/med_qa
43
+ - meta-math/MetaMathQA-40K
44
+ - openbookqa
45
+ - piqa
46
+ - metaeval/reclor
47
+ - derek-thomas/ScienceQA
48
+ - scibench
49
+ - sciq
50
+ - Open-Orca/SlimOrca
51
+ - migtissera/Synthia-v1.3
52
+ - TIGER-Lab/ScienceEval
53
+ model-index:
54
+ - name: Einstein-v4-7B
55
+ results:
56
+ - task:
57
+ type: text-generation
58
+ name: Text Generation
59
+ dataset:
60
+ name: AI2 Reasoning Challenge (25-Shot)
61
+ type: ai2_arc
62
+ config: ARC-Challenge
63
+ split: test
64
+ args:
65
+ num_few_shot: 25
66
+ metrics:
67
+ - type: acc_norm
68
+ value: 64.68
69
+ name: normalized accuracy
70
+ source:
71
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
72
+ name: Open LLM Leaderboard
73
+ - task:
74
+ type: text-generation
75
+ name: Text Generation
76
+ dataset:
77
+ name: HellaSwag (10-Shot)
78
+ type: hellaswag
79
+ split: validation
80
+ args:
81
+ num_few_shot: 10
82
+ metrics:
83
+ - type: acc_norm
84
+ value: 83.75
85
+ name: normalized accuracy
86
+ source:
87
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
88
+ name: Open LLM Leaderboard
89
+ - task:
90
+ type: text-generation
91
+ name: Text Generation
92
+ dataset:
93
+ name: MMLU (5-Shot)
94
+ type: cais/mmlu
95
+ config: all
96
+ split: test
97
+ args:
98
+ num_few_shot: 5
99
+ metrics:
100
+ - type: acc
101
+ value: 62.31
102
+ name: accuracy
103
+ source:
104
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
105
+ name: Open LLM Leaderboard
106
+ - task:
107
+ type: text-generation
108
+ name: Text Generation
109
+ dataset:
110
+ name: TruthfulQA (0-shot)
111
+ type: truthful_qa
112
+ config: multiple_choice
113
+ split: validation
114
+ args:
115
+ num_few_shot: 0
116
+ metrics:
117
+ - type: mc2
118
+ value: 55.15
119
+ source:
120
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
121
+ name: Open LLM Leaderboard
122
+ - task:
123
+ type: text-generation
124
+ name: Text Generation
125
+ dataset:
126
+ name: Winogrande (5-shot)
127
+ type: winogrande
128
+ config: winogrande_xl
129
+ split: validation
130
+ args:
131
+ num_few_shot: 5
132
+ metrics:
133
+ - type: acc
134
+ value: 76.24
135
+ name: accuracy
136
+ source:
137
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
138
+ name: Open LLM Leaderboard
139
+ - task:
140
+ type: text-generation
141
+ name: Text Generation
142
+ dataset:
143
+ name: GSM8k (5-shot)
144
+ type: gsm8k
145
+ config: main
146
+ split: test
147
+ args:
148
+ num_few_shot: 5
149
+ metrics:
150
+ - type: acc
151
+ value: 57.62
152
+ name: accuracy
153
+ source:
154
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
155
+ name: Open LLM Leaderboard
156
+ - task:
157
+ type: text-generation
158
+ name: Text Generation
159
+ dataset:
160
+ name: IFEval (0-Shot)
161
+ type: HuggingFaceH4/ifeval
162
+ args:
163
+ num_few_shot: 0
164
+ metrics:
165
+ - type: inst_level_strict_acc and prompt_level_strict_acc
166
+ value: 47.08
167
+ name: strict accuracy
168
+ source:
169
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
170
+ name: Open LLM Leaderboard
171
+ - task:
172
+ type: text-generation
173
+ name: Text Generation
174
+ dataset:
175
+ name: BBH (3-Shot)
176
+ type: BBH
177
+ args:
178
+ num_few_shot: 3
179
+ metrics:
180
+ - type: acc_norm
181
+ value: 14.3
182
+ name: normalized accuracy
183
+ source:
184
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
185
+ name: Open LLM Leaderboard
186
+ - task:
187
+ type: text-generation
188
+ name: Text Generation
189
+ dataset:
190
+ name: MATH Lvl 5 (4-Shot)
191
+ type: hendrycks/competition_math
192
+ args:
193
+ num_few_shot: 4
194
+ metrics:
195
+ - type: exact_match
196
+ value: 1.74
197
+ name: exact match
198
+ source:
199
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
200
+ name: Open LLM Leaderboard
201
+ - task:
202
+ type: text-generation
203
+ name: Text Generation
204
+ dataset:
205
+ name: GPQA (0-shot)
206
+ type: Idavidrein/gpqa
207
+ args:
208
+ num_few_shot: 0
209
+ metrics:
210
+ - type: acc_norm
211
+ value: 4.25
212
+ name: acc_norm
213
+ source:
214
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
215
+ name: Open LLM Leaderboard
216
+ - task:
217
+ type: text-generation
218
+ name: Text Generation
219
+ dataset:
220
+ name: MuSR (0-shot)
221
+ type: TAUR-Lab/MuSR
222
+ args:
223
+ num_few_shot: 0
224
+ metrics:
225
+ - type: acc_norm
226
+ value: 19.02
227
+ name: acc_norm
228
+ source:
229
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
230
+ name: Open LLM Leaderboard
231
+ - task:
232
+ type: text-generation
233
+ name: Text Generation
234
+ dataset:
235
+ name: MMLU-PRO (5-shot)
236
+ type: TIGER-Lab/MMLU-Pro
237
+ config: main
238
+ split: test
239
+ args:
240
+ num_few_shot: 5
241
+ metrics:
242
+ - type: acc
243
+ value: 13.99
244
+ name: accuracy
245
+ source:
246
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
247
+ name: Open LLM Leaderboard
248
+ ---
249
+
250
+ # AIronMind/Einstein-v4-7B-Q4_K_M-GGUF
251
+ This model was converted to GGUF format from [`Weyaxi/Einstein-v4-7B`](https://huggingface.co/Weyaxi/Einstein-v4-7B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
252
+ Refer to the [original model card](https://huggingface.co/Weyaxi/Einstein-v4-7B) for more details on the model.
253
+
254
+ ## Use with llama.cpp
255
+ Install llama.cpp through brew (works on Mac and Linux)
256
+
257
+ ```bash
258
+ brew install llama.cpp
259
+
260
+ ```
261
+ Invoke the llama.cpp server or the CLI.
262
+
263
+ ### CLI:
264
+ ```bash
265
+ llama-cli --hf-repo AIronMind/Einstein-v4-7B-Q4_K_M-GGUF --hf-file einstein-v4-7b-q4_k_m.gguf -p "The meaning to life and the universe is"
266
+ ```
267
+
268
+ ### Server:
269
+ ```bash
270
+ llama-server --hf-repo AIronMind/Einstein-v4-7B-Q4_K_M-GGUF --hf-file einstein-v4-7b-q4_k_m.gguf -c 2048
271
+ ```
272
+
273
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
274
+
275
+ Step 1: Clone llama.cpp from GitHub.
276
+ ```
277
+ git clone https://github.com/ggerganov/llama.cpp
278
+ ```
279
+
280
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
281
+ ```
282
+ cd llama.cpp && LLAMA_CURL=1 make
283
+ ```
284
+
285
+ Step 3: Run inference through the main binary.
286
+ ```
287
+ ./llama-cli --hf-repo AIronMind/Einstein-v4-7B-Q4_K_M-GGUF --hf-file einstein-v4-7b-q4_k_m.gguf -p "The meaning to life and the universe is"
288
+ ```
289
+ or
290
+ ```
291
+ ./llama-server --hf-repo AIronMind/Einstein-v4-7B-Q4_K_M-GGUF --hf-file einstein-v4-7b-q4_k_m.gguf -c 2048
292
+ ```