Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,7 @@
|
|
|
|
|
|
|
|
|
|
1 |
Reflection-70B is (currently) the world's top open-source LLM, trained with a new technique called Reflection-Tuning that teaches a LLM to detect mistakes in its reasoning and correct course.
|
2 |
|
3 |
The model was trained on synthetic data generated by [Glaive](https://glaive.ai). If you're training a model, Glaive is incredible — use them.
|
@@ -33,4 +37,8 @@ As mentioned above, the model uses the standard Llama 3.1 chat format. Here’s
|
|
33 |
You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside <thinking> tags, and then provide your final response inside <output> tags. If you detect that you made a mistake in your reasoning at any point, correct yourself inside <reflection> tags.<|eot_id|><|start_header_id|>user<|end_header_id|>
|
34 |
|
35 |
what is 2+2?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
|
36 |
-
```
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: llama3.1
|
3 |
+
base_model: meta-llama/Meta-Llama-3.1-70B-Instruct
|
4 |
+
---
|
5 |
Reflection-70B is (currently) the world's top open-source LLM, trained with a new technique called Reflection-Tuning that teaches a LLM to detect mistakes in its reasoning and correct course.
|
6 |
|
7 |
The model was trained on synthetic data generated by [Glaive](https://glaive.ai). If you're training a model, Glaive is incredible — use them.
|
|
|
37 |
You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside <thinking> tags, and then provide your final response inside <output> tags. If you detect that you made a mistake in your reasoning at any point, correct yourself inside <reflection> tags.<|eot_id|><|start_header_id|>user<|end_header_id|>
|
38 |
|
39 |
what is 2+2?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
|
40 |
+
```
|
41 |
+
|
42 |
+
## Dataset / Report
|
43 |
+
|
44 |
+
Both the dataset and a brief report detailing how we trained this model will be released next week, alongside out Reflection-405B model that we expect will be the top-performing LLM in the world, including closed-source models.
|