Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,36 @@
|
|
1 |
-
Reflection-70B is (currently) the world's top open-source LLM, trained with a new technique called
|
2 |
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Reflection-70B is (currently) the world's top open-source LLM, trained with a new technique called Reflection-Tuning that teaches a LLM to detect mistakes in its reasoning and correct course.
|
2 |
|
3 |
+
The model was trained on synthetic data generated by [Glaive](https://glaive.ai). If you're training a model, Glaive is incredible — use them.
|
4 |
+
|
5 |
+
## Benchmarks
|
6 |
+

|
7 |
+
|
8 |
+
Trained from Llama 3.1 70B Instruct, you can sample from Reflection-70B using the same code, pipelines, etc. as any other Llama model. It even uses the stock Llama 3.1 chat template format (though, we've trained in a few new special tokens to aid in reasoning and reflection).
|
9 |
+
|
10 |
+
During sampling, the model will start by outputting reasoning inside `<thinking>` and `</thinking>` tags, and then once it is satisfied with its reasoning, it will output the final answer inside `<output>` and `</output>` tags. Each of these tags are special tokens, trained into the model.
|
11 |
+
|
12 |
+
This enables the model to separate its internal thoughts and reasoning from its final answer, improving the experience for the user.
|
13 |
+
|
14 |
+
Inside the `<thinking>` section, the model may output one or more `<reflection>` tags, which signals the model has caught an error in its reasoning and will attempt to correct it before providing a final answer.
|
15 |
+
|
16 |
+
## System Prompt
|
17 |
+
|
18 |
+
```
|
19 |
+
The system prompt used for training this model is:
|
20 |
+
|
21 |
+
You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside <thinking> tags, and then provide your final response inside <output> tags. If you detect that you made a mistake in your reasoning at any point, correct yourself inside <reflection> tags.
|
22 |
+
|
23 |
+
We recommend using this exact system prompt to get the best results from Reflection-70B. You may also want to experiment combining this system prompt with your own custom instructions to customize the behavior of the model.
|
24 |
+
```
|
25 |
+
|
26 |
+
## Chat Format
|
27 |
+
|
28 |
+
As mentioned above, the model uses the standard Llama 3.1 chat format. Here’s an example:
|
29 |
+
|
30 |
+
```
|
31 |
+
<|begin_of_text|><|start_header_id|>system<|end_header_id|>
|
32 |
+
|
33 |
+
You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside <thinking> tags, and then provide your final response inside <output> tags. If you detect that you made a mistake in your reasoning at any point, correct yourself inside <reflection> tags.<|eot_id|><|start_header_id|>user<|end_header_id|>
|
34 |
+
|
35 |
+
what is 2+2?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
|
36 |
+
```
|