Updated `lmsys-chat-1m-synth-{ja,en}-wo-pii-and-template-instructions`
Browse files
README.md
CHANGED
@@ -177,20 +177,21 @@ print(output[0].outputs[0].text)
|
|
177 |
|
178 |
### Instruction Tuning
|
179 |
|
180 |
-
The following datasets were used for the instruction tuning.
|
181 |
-
|
182 |
-
- lmsys-chat-1m-synth-ja-wo-pii
|
183 |
-
|
184 |
-
-
|
185 |
-
|
186 |
-
|
187 |
-
- filtered magpie-ultra
|
188 |
- Subset of the [magpie-ultra](https://huggingface.co/datasets/argilla/magpie-ultra-v0.1) dataset, containing samples rated as 'average,' 'good,' or 'excellent.'.
|
189 |
-
|
190 |
-
- gemma-magpie
|
191 |
-
- Japanese dataset.
|
192 |
- Generated using prompts for specific category words.
|
193 |
-
|
|
|
|
|
|
|
|
|
|
|
194 |
## Risks and Limitations
|
195 |
|
196 |
The models released here are still in the early stages of our research and development and have not been tuned to ensure outputs align with human intent and safety considerations.
|
|
|
177 |
|
178 |
### Instruction Tuning
|
179 |
|
180 |
+
The following instruction datasets were used for the instruction tuning.
|
181 |
+
- Japanese
|
182 |
+
- `lmsys-chat-1m-synth-ja-wo-pii-and-template-instructions`
|
183 |
+
- Single-turn Japanese synthetic instruction dataset derived from [lmsys-chat-1m](https://huggingface.co/datasets/lmsys/lmsys-chat-1m) dataset [\[Zhang+, ICLR24\]](https://openreview.net/forum?id=BOfDKxfwt0)). The user instructions were translated into Japanese via machine translation, and the assistant responses were generated using the Llama 3.1 405B Instruct model. Rejection sampling was applied, with Llama 3.1 70B Instruct serving as the judge.
|
184 |
+
- As implied by the dataset name, conversations that contain personally identifiable information (PII) or template-based user instructions have been removed.
|
185 |
+
- filtered magpie-ultra-ja
|
|
|
|
|
186 |
- Subset of the [magpie-ultra](https://huggingface.co/datasets/argilla/magpie-ultra-v0.1) dataset, containing samples rated as 'average,' 'good,' or 'excellent.'.
|
187 |
+
- gemma-magpie
|
|
|
|
|
188 |
- Generated using prompts for specific category words.
|
189 |
+
- English
|
190 |
+
- `lmsys-chat-1m-synth-en-wo-pii-and-template-instructions`
|
191 |
+
- Similar to the `lmsys-chat-1m-synth-ja-wo-pii-and-template-instructions`, but this version uses the original English user instructions. The assistant responses were generated in English as well. Rejection sampling was not applied in this version.
|
192 |
+
- filtered-magpie-ultra-en
|
193 |
+
- Subset of the [magpie-ultra](https://huggingface.co/datasets/argilla/magpie-ultra-v0.1) dataset, containing samples rated as 'average,' 'good,' or 'excellent.'.
|
194 |
+
|
195 |
## Risks and Limitations
|
196 |
|
197 |
The models released here are still in the early stages of our research and development and have not been tuned to ensure outputs align with human intent and safety considerations.
|