mattshumer commited on
Commit
fb47a3a
·
verified ·
1 Parent(s): 4f250d6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -1
README.md CHANGED
@@ -1,3 +1,7 @@
 
 
 
 
1
  Reflection-70B is (currently) the world's top open-source LLM, trained with a new technique called Reflection-Tuning that teaches a LLM to detect mistakes in its reasoning and correct course.
2
 
3
  The model was trained on synthetic data generated by [Glaive](https://glaive.ai). If you're training a model, Glaive is incredible — use them.
@@ -33,4 +37,8 @@ As mentioned above, the model uses the standard Llama 3.1 chat format. Here’s
33
  You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside <thinking> tags, and then provide your final response inside <output> tags. If you detect that you made a mistake in your reasoning at any point, correct yourself inside <reflection> tags.<|eot_id|><|start_header_id|>user<|end_header_id|>
34
 
35
  what is 2+2?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
36
- ```
 
 
 
 
 
1
+ ---
2
+ license: llama3.1
3
+ base_model: meta-llama/Meta-Llama-3.1-70B-Instruct
4
+ ---
5
  Reflection-70B is (currently) the world's top open-source LLM, trained with a new technique called Reflection-Tuning that teaches a LLM to detect mistakes in its reasoning and correct course.
6
 
7
  The model was trained on synthetic data generated by [Glaive](https://glaive.ai). If you're training a model, Glaive is incredible — use them.
 
37
  You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside <thinking> tags, and then provide your final response inside <output> tags. If you detect that you made a mistake in your reasoning at any point, correct yourself inside <reflection> tags.<|eot_id|><|start_header_id|>user<|end_header_id|>
38
 
39
  what is 2+2?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
40
+ ```
41
+
42
+ ## Dataset / Report
43
+
44
+ Both the dataset and a brief report detailing how we trained this model will be released next week, alongside out Reflection-405B model that we expect will be the top-performing LLM in the world, including closed-source models.