Update README.md
Browse files
README.md
CHANGED
@@ -1,7 +1,5 @@
|
|
1 |
-
|
2 |
-
|
3 |
-
base_model: meta-llama/Meta-Llama-3.1-70B-Instruct
|
4 |
-
---
|
5 |
**Reflection 70B is (currently) the world's top open-source LLM, trained with a new technique called Reflection-Tuning that teaches a LLM to detect mistakes in its reasoning and correct course.**
|
6 |
|
7 |
The model was trained on synthetic data generated by [Glaive](https://glaive.ai). If you're training a model, Glaive is incredible — use them.
|
@@ -39,10 +37,20 @@ You are a world-class AI system, capable of complex reasoning and reflection. Re
|
|
39 |
what is 2+2?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
|
40 |
```
|
41 |
|
|
|
|
|
|
|
|
|
|
|
42 |
## Dataset / Report
|
43 |
|
44 |
Both the dataset and a brief report detailing how we trained this model will be released next week, alongside our Reflection 405B model that we expect will be the top-performing LLM in the world, including closed-source models.
|
45 |
|
46 |
---
|
47 |
|
48 |
-
Thanks to Jason Kuperberg and Josh Bickett from the [HyperWrite](https://hyperwriteai.com) team for reviewing drafts of the report we'll be releasing next week.
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Reflection 70B
|
2 |
+
|
|
|
|
|
3 |
**Reflection 70B is (currently) the world's top open-source LLM, trained with a new technique called Reflection-Tuning that teaches a LLM to detect mistakes in its reasoning and correct course.**
|
4 |
|
5 |
The model was trained on synthetic data generated by [Glaive](https://glaive.ai). If you're training a model, Glaive is incredible — use them.
|
|
|
37 |
what is 2+2?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
|
38 |
```
|
39 |
|
40 |
+
## Tips for Performance
|
41 |
+
|
42 |
+
- We are initially recommending a `temperature` of `.7` and a `top_p` of `.95`.
|
43 |
+
- For increased accuracy, append `Think carefully.` at the end of your messages.
|
44 |
+
|
45 |
## Dataset / Report
|
46 |
|
47 |
Both the dataset and a brief report detailing how we trained this model will be released next week, alongside our Reflection 405B model that we expect will be the top-performing LLM in the world, including closed-source models.
|
48 |
|
49 |
---
|
50 |
|
51 |
+
Thanks to Jason Kuperberg and Josh Bickett from the [HyperWrite](https://hyperwriteai.com) team for reviewing drafts of the report we'll be releasing next week.
|
52 |
+
|
53 |
+
---
|
54 |
+
license: llama3.1
|
55 |
+
base_model: meta-llama/Meta-Llama-3.1-70B-Instruct
|
56 |
+
---
|