mattshumer commited on
Commit
d1a95bc
·
verified ·
1 Parent(s): da0d153

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -2,14 +2,14 @@
2
  license: llama3.1
3
  base_model: meta-llama/Meta-Llama-3.1-70B-Instruct
4
  ---
5
- Reflection-70B is (currently) the world's top open-source LLM, trained with a new technique called Reflection-Tuning that teaches a LLM to detect mistakes in its reasoning and correct course.
6
 
7
  The model was trained on synthetic data generated by [Glaive](https://glaive.ai). If you're training a model, Glaive is incredible — use them.
8
 
9
  ## Benchmarks
10
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/60518f3731c5be7f3dd5ebc3/QPqof1rVoT88qmif_9LSu.png)
11
 
12
- Trained from Llama 3.1 70B Instruct, you can sample from Reflection-70B using the same code, pipelines, etc. as any other Llama model. It even uses the stock Llama 3.1 chat template format (though, we've trained in a few new special tokens to aid in reasoning and reflection).
13
 
14
  During sampling, the model will start by outputting reasoning inside `<thinking>` and `</thinking>` tags, and then once it is satisfied with its reasoning, it will output the final answer inside `<output>` and `</output>` tags. Each of these tags are special tokens, trained into the model.
15
 
@@ -24,7 +24,7 @@ The system prompt used for training this model is:
24
 
25
  You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside <thinking> tags, and then provide your final response inside <output> tags. If you detect that you made a mistake in your reasoning at any point, correct yourself inside <reflection> tags.
26
 
27
- We recommend using this exact system prompt to get the best results from Reflection-70B. You may also want to experiment combining this system prompt with your own custom instructions to customize the behavior of the model.
28
  ```
29
 
30
  ## Chat Format
@@ -41,7 +41,7 @@ what is 2+2?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
41
 
42
  ## Dataset / Report
43
 
44
- Both the dataset and a brief report detailing how we trained this model will be released next week, alongside out Reflection-405B model that we expect will be the top-performing LLM in the world, including closed-source models.
45
 
46
  ---
47
 
 
2
  license: llama3.1
3
  base_model: meta-llama/Meta-Llama-3.1-70B-Instruct
4
  ---
5
+ **Reflection 70B is (currently) the world's top open-source LLM, trained with a new technique called Reflection-Tuning that teaches a LLM to detect mistakes in its reasoning and correct course.**
6
 
7
  The model was trained on synthetic data generated by [Glaive](https://glaive.ai). If you're training a model, Glaive is incredible — use them.
8
 
9
  ## Benchmarks
10
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/60518f3731c5be7f3dd5ebc3/QPqof1rVoT88qmif_9LSu.png)
11
 
12
+ Trained from Llama 3.1 70B Instruct, you can sample from Reflection 70B using the same code, pipelines, etc. as any other Llama model. It even uses the stock Llama 3.1 chat template format (though, we've trained in a few new special tokens to aid in reasoning and reflection).
13
 
14
  During sampling, the model will start by outputting reasoning inside `<thinking>` and `</thinking>` tags, and then once it is satisfied with its reasoning, it will output the final answer inside `<output>` and `</output>` tags. Each of these tags are special tokens, trained into the model.
15
 
 
24
 
25
  You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside <thinking> tags, and then provide your final response inside <output> tags. If you detect that you made a mistake in your reasoning at any point, correct yourself inside <reflection> tags.
26
 
27
+ We recommend using this exact system prompt to get the best results from Reflection 70B. You may also want to experiment combining this system prompt with your own custom instructions to customize the behavior of the model.
28
  ```
29
 
30
  ## Chat Format
 
41
 
42
  ## Dataset / Report
43
 
44
+ Both the dataset and a brief report detailing how we trained this model will be released next week, alongside our Reflection 405B model that we expect will be the top-performing LLM in the world, including closed-source models.
45
 
46
  ---
47