mattshumer commited on
Commit
4f250d6
·
verified ·
1 Parent(s): afe57db

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +35 -2
README.md CHANGED
@@ -1,3 +1,36 @@
1
- Reflection-70B is (currently) the world's top open-source LLM, trained with a new technique called `Reflection-Tuning` that teaches a LLM to detect mistakes in its reasoning and correct course.
2
 
3
- Trained from Llama 3.1 70B Instruct, you can sample from Reflection-70B using the same code, pipelines, etc. as Llama
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Reflection-70B is (currently) the world's top open-source LLM, trained with a new technique called Reflection-Tuning that teaches a LLM to detect mistakes in its reasoning and correct course.
2
 
3
+ The model was trained on synthetic data generated by [Glaive](https://glaive.ai). If you're training a model, Glaive is incredible — use them.
4
+
5
+ ## Benchmarks
6
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/60518f3731c5be7f3dd5ebc3/QPqof1rVoT88qmif_9LSu.png)
7
+
8
+ Trained from Llama 3.1 70B Instruct, you can sample from Reflection-70B using the same code, pipelines, etc. as any other Llama model. It even uses the stock Llama 3.1 chat template format (though, we've trained in a few new special tokens to aid in reasoning and reflection).
9
+
10
+ During sampling, the model will start by outputting reasoning inside `<thinking>` and `</thinking>` tags, and then once it is satisfied with its reasoning, it will output the final answer inside `<output>` and `</output>` tags. Each of these tags are special tokens, trained into the model.
11
+
12
+ This enables the model to separate its internal thoughts and reasoning from its final answer, improving the experience for the user.
13
+
14
+ Inside the `<thinking>` section, the model may output one or more `<reflection>` tags, which signals the model has caught an error in its reasoning and will attempt to correct it before providing a final answer.
15
+
16
+ ## System Prompt
17
+
18
+ ```
19
+ The system prompt used for training this model is:
20
+
21
+ You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside <thinking> tags, and then provide your final response inside <output> tags. If you detect that you made a mistake in your reasoning at any point, correct yourself inside <reflection> tags.
22
+
23
+ We recommend using this exact system prompt to get the best results from Reflection-70B. You may also want to experiment combining this system prompt with your own custom instructions to customize the behavior of the model.
24
+ ```
25
+
26
+ ## Chat Format
27
+
28
+ As mentioned above, the model uses the standard Llama 3.1 chat format. Here’s an example:
29
+
30
+ ```
31
+ <|begin_of_text|><|start_header_id|>system<|end_header_id|>
32
+
33
+ You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside <thinking> tags, and then provide your final response inside <output> tags. If you detect that you made a mistake in your reasoning at any point, correct yourself inside <reflection> tags.<|eot_id|><|start_header_id|>user<|end_header_id|>
34
+
35
+ what is 2+2?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
36
+ ```