Update README.md
Browse files
README.md
CHANGED
@@ -3,17 +3,19 @@ inference: false
|
|
3 |
license: other
|
4 |
---
|
5 |
|
|
|
6 |
<div style="width: 100%;">
|
7 |
<img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
|
8 |
</div>
|
9 |
<div style="display: flex; justify-content: space-between; width: 100%;">
|
10 |
<div style="display: flex; flex-direction: column; align-items: flex-start;">
|
11 |
-
<p><a href="https://discord.gg/
|
12 |
</div>
|
13 |
<div style="display: flex; flex-direction: column; align-items: flex-end;">
|
14 |
-
<p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon
|
15 |
</div>
|
16 |
</div>
|
|
|
17 |
|
18 |
# WizardLM 13B 1.0 fp16
|
19 |
|
@@ -39,18 +41,31 @@ USER: prompt goes here
|
|
39 |
ASSISTANT:
|
40 |
```
|
41 |
|
42 |
-
|
|
|
43 |
|
44 |
-
|
45 |
|
46 |
-
|
47 |
|
48 |
-
|
49 |
|
50 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
51 |
* Ko-Fi: https://ko-fi.com/TheBlokeAI
|
52 |
-
|
53 |
-
|
|
|
|
|
|
|
|
|
54 |
# Original model card
|
55 |
|
56 |
## WizardLM: An Instruction-following LLM Using Evol-Instruct
|
@@ -80,7 +95,7 @@ At present, our core contributors are preparing the **33B** version and we expec
|
|
80 |
|
81 |
### GPT-4 automatic evaluation
|
82 |
|
83 |
-
We adopt the automatic evaluation framework based on GPT-4 proposed by FastChat to assess the performance of chatbot models. As shown in the following figure, WizardLM-13B achieved better results than Vicuna-13b.
|
84 |
<p align="center" width="100%">
|
85 |
<a ><img src="imgs/WizarLM13b-GPT4.png" alt="WizardLM" style="width: 100%; min-width: 300px; display: block; margin: auto;"></a>
|
86 |
</p>
|
@@ -161,7 +176,7 @@ This JSON file is a list of dictionaries, each dictionary contains the following
|
|
161 |
We release [WizardLM] weights as delta weights to comply with the LLaMA model license.
|
162 |
You can add our delta to the original LLaMA weights to obtain the WizardLM weights. Instructions:
|
163 |
1. Get the original LLaMA weights in the huggingface format by following the instructions [here](https://huggingface.co/docs/transformers/main/model_doc/llama).
|
164 |
-
2. Please download our delta model at the following [link](https://huggingface.co/victor123/WizardLM)
|
165 |
3. Use the following scripts to get WizardLM weights by applying our delta:
|
166 |
```
|
167 |
python src/weight_diff_wizard.py recover --path_raw <path_to_step_1_dir> --path_diff <path_to_step_2_dir> --path_tuned <path_to_store_recovered_weights>
|
@@ -224,9 +239,9 @@ python src\inference_wizardlm.py
|
|
224 |
|
225 |
### Evaluation
|
226 |
|
227 |
-
To evaluate Wizard, we conduct human evaluation on the inputs from our human instruct evaluation set [`WizardLM_testset.jsonl`](./data/WizardLM_testset.jsonl) . This evaluation set was collected by the authors and covers a diverse list of user-oriented instructions including difficult Coding Generation & Debugging, Math, Reasoning, Complex Formats, Academic Writing, Extensive Disciplines, and so on. We performed a blind pairwise comparison between Wizard and baselines. Specifically, we recruit 10 well-educated annotators to rank the models from 1 to 5 on relevance, knowledgeable, reasoning, calculation and accuracy.
|
228 |
|
229 |
-
WizardLM achieved significantly better results than Alpaca and Vicuna-7b.
|
230 |
<p align="center" width="60%">
|
231 |
<a ><img src="imgs/win.png" alt="WizardLM" style="width: 60%; min-width: 300px; display: block; margin: auto;"></a>
|
232 |
</p>
|
@@ -242,7 +257,7 @@ Please cite the repo if you use the data or code in this repo.
|
|
242 |
|
243 |
```
|
244 |
@misc{xu2023wizardlm,
|
245 |
-
title={WizardLM: Empowering Large Language Models to Follow Complex Instructions},
|
246 |
author={Can Xu and Qingfeng Sun and Kai Zheng and Xiubo Geng and Pu Zhao and Jiazhan Feng and Chongyang Tao and Daxin Jiang},
|
247 |
year={2023},
|
248 |
eprint={2304.12244},
|
|
|
3 |
license: other
|
4 |
---
|
5 |
|
6 |
+
<!-- header start -->
|
7 |
<div style="width: 100%;">
|
8 |
<img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
|
9 |
</div>
|
10 |
<div style="display: flex; justify-content: space-between; width: 100%;">
|
11 |
<div style="display: flex; flex-direction: column; align-items: flex-start;">
|
12 |
+
<p><a href="https://discord.gg/Jq4vkcDakD">Chat & support: my new Discord server</a></p>
|
13 |
</div>
|
14 |
<div style="display: flex; flex-direction: column; align-items: flex-end;">
|
15 |
+
<p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p>
|
16 |
</div>
|
17 |
</div>
|
18 |
+
<!-- header end -->
|
19 |
|
20 |
# WizardLM 13B 1.0 fp16
|
21 |
|
|
|
41 |
ASSISTANT:
|
42 |
```
|
43 |
|
44 |
+
<!-- footer start -->
|
45 |
+
## Discord
|
46 |
|
47 |
+
For further support, and discussions on these models and AI in general, join us at:
|
48 |
|
49 |
+
[TheBloke AI's Discord server](https://discord.gg/Jq4vkcDakD)
|
50 |
|
51 |
+
## Thanks, and how to contribute.
|
52 |
|
53 |
+
Thanks to the [chirper.ai](https://chirper.ai) team!
|
54 |
+
|
55 |
+
I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training.
|
56 |
+
|
57 |
+
If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects.
|
58 |
+
|
59 |
+
Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits.
|
60 |
+
|
61 |
+
* Patreon: https://patreon.com/TheBlokeAI
|
62 |
* Ko-Fi: https://ko-fi.com/TheBlokeAI
|
63 |
+
|
64 |
+
**Patreon special mentions**: Aemon Algiz, Dmitriy Samsonov, Nathan LeClaire, Trenton Dambrowitz, Mano Prime, David Flickinger, vamX, Nikolai Manek, senxiiz, Khalefa Al-Ahmad, Illia Dulskyi, Jonathan Leane, Talal Aujan, V. Lukas, Joseph William Delisle, Pyrater, Oscar Rangel, Lone Striker, Luke Pendergrass, Eugene Pentland, Sebastain Graf, Johann-Peter Hartman.
|
65 |
+
|
66 |
+
Thank you to all my generous patrons and donaters!
|
67 |
+
<!-- footer end -->
|
68 |
+
|
69 |
# Original model card
|
70 |
|
71 |
## WizardLM: An Instruction-following LLM Using Evol-Instruct
|
|
|
95 |
|
96 |
### GPT-4 automatic evaluation
|
97 |
|
98 |
+
We adopt the automatic evaluation framework based on GPT-4 proposed by FastChat to assess the performance of chatbot models. As shown in the following figure, WizardLM-13B achieved better results than Vicuna-13b.
|
99 |
<p align="center" width="100%">
|
100 |
<a ><img src="imgs/WizarLM13b-GPT4.png" alt="WizardLM" style="width: 100%; min-width: 300px; display: block; margin: auto;"></a>
|
101 |
</p>
|
|
|
176 |
We release [WizardLM] weights as delta weights to comply with the LLaMA model license.
|
177 |
You can add our delta to the original LLaMA weights to obtain the WizardLM weights. Instructions:
|
178 |
1. Get the original LLaMA weights in the huggingface format by following the instructions [here](https://huggingface.co/docs/transformers/main/model_doc/llama).
|
179 |
+
2. Please download our delta model at the following [link](https://huggingface.co/victor123/WizardLM)
|
180 |
3. Use the following scripts to get WizardLM weights by applying our delta:
|
181 |
```
|
182 |
python src/weight_diff_wizard.py recover --path_raw <path_to_step_1_dir> --path_diff <path_to_step_2_dir> --path_tuned <path_to_store_recovered_weights>
|
|
|
239 |
|
240 |
### Evaluation
|
241 |
|
242 |
+
To evaluate Wizard, we conduct human evaluation on the inputs from our human instruct evaluation set [`WizardLM_testset.jsonl`](./data/WizardLM_testset.jsonl) . This evaluation set was collected by the authors and covers a diverse list of user-oriented instructions including difficult Coding Generation & Debugging, Math, Reasoning, Complex Formats, Academic Writing, Extensive Disciplines, and so on. We performed a blind pairwise comparison between Wizard and baselines. Specifically, we recruit 10 well-educated annotators to rank the models from 1 to 5 on relevance, knowledgeable, reasoning, calculation and accuracy.
|
243 |
|
244 |
+
WizardLM achieved significantly better results than Alpaca and Vicuna-7b.
|
245 |
<p align="center" width="60%">
|
246 |
<a ><img src="imgs/win.png" alt="WizardLM" style="width: 60%; min-width: 300px; display: block; margin: auto;"></a>
|
247 |
</p>
|
|
|
257 |
|
258 |
```
|
259 |
@misc{xu2023wizardlm,
|
260 |
+
title={WizardLM: Empowering Large Language Models to Follow Complex Instructions},
|
261 |
author={Can Xu and Qingfeng Sun and Kai Zheng and Xiubo Geng and Pu Zhao and Jiazhan Feng and Chongyang Tao and Daxin Jiang},
|
262 |
year={2023},
|
263 |
eprint={2304.12244},
|