TheBloke commited on
Commit
b797338
1 Parent(s): a711f0b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +29 -14
README.md CHANGED
@@ -3,17 +3,19 @@ inference: false
3
  license: other
4
  ---
5
 
 
6
  <div style="width: 100%;">
7
  <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
8
  </div>
9
  <div style="display: flex; justify-content: space-between; width: 100%;">
10
  <div style="display: flex; flex-direction: column; align-items: flex-start;">
11
- <p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
12
  </div>
13
  <div style="display: flex; flex-direction: column; align-items: flex-end;">
14
- <p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon coming soon!</a></p>
15
  </div>
16
  </div>
 
17
 
18
  # WizardLM 13B 1.0 fp16
19
 
@@ -39,18 +41,31 @@ USER: prompt goes here
39
  ASSISTANT:
40
  ```
41
 
42
- ## Want to support my work?
 
43
 
44
- I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills.
45
 
46
- So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects.
47
 
48
- Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try.
49
 
50
- * Patreon: coming soon! (just awaiting approval)
 
 
 
 
 
 
 
 
51
  * Ko-Fi: https://ko-fi.com/TheBlokeAI
52
- * Discord: https://discord.gg/UBgz4VXf
53
-
 
 
 
 
54
  # Original model card
55
 
56
  ## WizardLM: An Instruction-following LLM Using Evol-Instruct
@@ -80,7 +95,7 @@ At present, our core contributors are preparing the **33B** version and we expec
80
 
81
  ### GPT-4 automatic evaluation
82
 
83
- We adopt the automatic evaluation framework based on GPT-4 proposed by FastChat to assess the performance of chatbot models. As shown in the following figure, WizardLM-13B achieved better results than Vicuna-13b.
84
  <p align="center" width="100%">
85
  <a ><img src="imgs/WizarLM13b-GPT4.png" alt="WizardLM" style="width: 100%; min-width: 300px; display: block; margin: auto;"></a>
86
  </p>
@@ -161,7 +176,7 @@ This JSON file is a list of dictionaries, each dictionary contains the following
161
  We release [WizardLM] weights as delta weights to comply with the LLaMA model license.
162
  You can add our delta to the original LLaMA weights to obtain the WizardLM weights. Instructions:
163
  1. Get the original LLaMA weights in the huggingface format by following the instructions [here](https://huggingface.co/docs/transformers/main/model_doc/llama).
164
- 2. Please download our delta model at the following [link](https://huggingface.co/victor123/WizardLM)
165
  3. Use the following scripts to get WizardLM weights by applying our delta:
166
  ```
167
  python src/weight_diff_wizard.py recover --path_raw <path_to_step_1_dir> --path_diff <path_to_step_2_dir> --path_tuned <path_to_store_recovered_weights>
@@ -224,9 +239,9 @@ python src\inference_wizardlm.py
224
 
225
  ### Evaluation
226
 
227
- To evaluate Wizard, we conduct human evaluation on the inputs from our human instruct evaluation set [`WizardLM_testset.jsonl`](./data/WizardLM_testset.jsonl) . This evaluation set was collected by the authors and covers a diverse list of user-oriented instructions including difficult Coding Generation & Debugging, Math, Reasoning, Complex Formats, Academic Writing, Extensive Disciplines, and so on. We performed a blind pairwise comparison between Wizard and baselines. Specifically, we recruit 10 well-educated annotators to rank the models from 1 to 5 on relevance, knowledgeable, reasoning, calculation and accuracy.
228
 
229
- WizardLM achieved significantly better results than Alpaca and Vicuna-7b.
230
  <p align="center" width="60%">
231
  <a ><img src="imgs/win.png" alt="WizardLM" style="width: 60%; min-width: 300px; display: block; margin: auto;"></a>
232
  </p>
@@ -242,7 +257,7 @@ Please cite the repo if you use the data or code in this repo.
242
 
243
  ```
244
  @misc{xu2023wizardlm,
245
- title={WizardLM: Empowering Large Language Models to Follow Complex Instructions},
246
  author={Can Xu and Qingfeng Sun and Kai Zheng and Xiubo Geng and Pu Zhao and Jiazhan Feng and Chongyang Tao and Daxin Jiang},
247
  year={2023},
248
  eprint={2304.12244},
 
3
  license: other
4
  ---
5
 
6
+ <!-- header start -->
7
  <div style="width: 100%;">
8
  <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
9
  </div>
10
  <div style="display: flex; justify-content: space-between; width: 100%;">
11
  <div style="display: flex; flex-direction: column; align-items: flex-start;">
12
+ <p><a href="https://discord.gg/Jq4vkcDakD">Chat & support: my new Discord server</a></p>
13
  </div>
14
  <div style="display: flex; flex-direction: column; align-items: flex-end;">
15
+ <p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p>
16
  </div>
17
  </div>
18
+ <!-- header end -->
19
 
20
  # WizardLM 13B 1.0 fp16
21
 
 
41
  ASSISTANT:
42
  ```
43
 
44
+ <!-- footer start -->
45
+ ## Discord
46
 
47
+ For further support, and discussions on these models and AI in general, join us at:
48
 
49
+ [TheBloke AI's Discord server](https://discord.gg/Jq4vkcDakD)
50
 
51
+ ## Thanks, and how to contribute.
52
 
53
+ Thanks to the [chirper.ai](https://chirper.ai) team!
54
+
55
+ I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training.
56
+
57
+ If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects.
58
+
59
+ Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits.
60
+
61
+ * Patreon: https://patreon.com/TheBlokeAI
62
  * Ko-Fi: https://ko-fi.com/TheBlokeAI
63
+
64
+ **Patreon special mentions**: Aemon Algiz, Dmitriy Samsonov, Nathan LeClaire, Trenton Dambrowitz, Mano Prime, David Flickinger, vamX, Nikolai Manek, senxiiz, Khalefa Al-Ahmad, Illia Dulskyi, Jonathan Leane, Talal Aujan, V. Lukas, Joseph William Delisle, Pyrater, Oscar Rangel, Lone Striker, Luke Pendergrass, Eugene Pentland, Sebastain Graf, Johann-Peter Hartman.
65
+
66
+ Thank you to all my generous patrons and donaters!
67
+ <!-- footer end -->
68
+
69
  # Original model card
70
 
71
  ## WizardLM: An Instruction-following LLM Using Evol-Instruct
 
95
 
96
  ### GPT-4 automatic evaluation
97
 
98
+ We adopt the automatic evaluation framework based on GPT-4 proposed by FastChat to assess the performance of chatbot models. As shown in the following figure, WizardLM-13B achieved better results than Vicuna-13b.
99
  <p align="center" width="100%">
100
  <a ><img src="imgs/WizarLM13b-GPT4.png" alt="WizardLM" style="width: 100%; min-width: 300px; display: block; margin: auto;"></a>
101
  </p>
 
176
  We release [WizardLM] weights as delta weights to comply with the LLaMA model license.
177
  You can add our delta to the original LLaMA weights to obtain the WizardLM weights. Instructions:
178
  1. Get the original LLaMA weights in the huggingface format by following the instructions [here](https://huggingface.co/docs/transformers/main/model_doc/llama).
179
+ 2. Please download our delta model at the following [link](https://huggingface.co/victor123/WizardLM)
180
  3. Use the following scripts to get WizardLM weights by applying our delta:
181
  ```
182
  python src/weight_diff_wizard.py recover --path_raw <path_to_step_1_dir> --path_diff <path_to_step_2_dir> --path_tuned <path_to_store_recovered_weights>
 
239
 
240
  ### Evaluation
241
 
242
+ To evaluate Wizard, we conduct human evaluation on the inputs from our human instruct evaluation set [`WizardLM_testset.jsonl`](./data/WizardLM_testset.jsonl) . This evaluation set was collected by the authors and covers a diverse list of user-oriented instructions including difficult Coding Generation & Debugging, Math, Reasoning, Complex Formats, Academic Writing, Extensive Disciplines, and so on. We performed a blind pairwise comparison between Wizard and baselines. Specifically, we recruit 10 well-educated annotators to rank the models from 1 to 5 on relevance, knowledgeable, reasoning, calculation and accuracy.
243
 
244
+ WizardLM achieved significantly better results than Alpaca and Vicuna-7b.
245
  <p align="center" width="60%">
246
  <a ><img src="imgs/win.png" alt="WizardLM" style="width: 60%; min-width: 300px; display: block; margin: auto;"></a>
247
  </p>
 
257
 
258
  ```
259
  @misc{xu2023wizardlm,
260
+ title={WizardLM: Empowering Large Language Models to Follow Complex Instructions},
261
  author={Can Xu and Qingfeng Sun and Kai Zheng and Xiubo Geng and Pu Zhao and Jiazhan Feng and Chongyang Tao and Daxin Jiang},
262
  year={2023},
263
  eprint={2304.12244},