Text Generation
Transformers
Safetensors
English
Japanese
llama
conversational
text-generation-inference
s-mizuki-nlp commited on
Commit
c09e444
·
verified ·
1 Parent(s): e6b649e

Updated `lmsys-chat-1m-synth-{ja,en}-wo-pii-and-template-instructions`

Browse files
Files changed (1) hide show
  1. README.md +13 -12
README.md CHANGED
@@ -177,20 +177,21 @@ print(output[0].outputs[0].text)
177
 
178
  ### Instruction Tuning
179
 
180
- The following datasets were used for the instruction tuning.
181
-
182
- - lmsys-chat-1m-synth-ja-wo-pii
183
-
184
- - Japanese translation of the lmsys-chat-1m dataset using DeepL, with synthetic instruction data created using the Llama-3.1-405B model.
185
- - 'wo-pii' indicates removal of personally identifiable information.
186
-
187
- - filtered magpie-ultra
188
  - Subset of the [magpie-ultra](https://huggingface.co/datasets/argilla/magpie-ultra-v0.1) dataset, containing samples rated as 'average,' 'good,' or 'excellent.'.
189
-
190
- - gemma-magpie
191
- - Japanese dataset.
192
  - Generated using prompts for specific category words.
193
-
 
 
 
 
 
194
  ## Risks and Limitations
195
 
196
  The models released here are still in the early stages of our research and development and have not been tuned to ensure outputs align with human intent and safety considerations.
 
177
 
178
  ### Instruction Tuning
179
 
180
+ The following instruction datasets were used for the instruction tuning.
181
+ - Japanese
182
+ - `lmsys-chat-1m-synth-ja-wo-pii-and-template-instructions`
183
+ - Single-turn Japanese synthetic instruction dataset derived from [lmsys-chat-1m](https://huggingface.co/datasets/lmsys/lmsys-chat-1m) dataset [\[Zhang+, ICLR24\]](https://openreview.net/forum?id=BOfDKxfwt0)). The user instructions were translated into Japanese via machine translation, and the assistant responses were generated using the Llama 3.1 405B Instruct model. Rejection sampling was applied, with Llama 3.1 70B Instruct serving as the judge.
184
+ - As implied by the dataset name, conversations that contain personally identifiable information (PII) or template-based user instructions have been removed.
185
+ - filtered magpie-ultra-ja
 
 
186
  - Subset of the [magpie-ultra](https://huggingface.co/datasets/argilla/magpie-ultra-v0.1) dataset, containing samples rated as 'average,' 'good,' or 'excellent.'.
187
+ - gemma-magpie
 
 
188
  - Generated using prompts for specific category words.
189
+ - English
190
+ - `lmsys-chat-1m-synth-en-wo-pii-and-template-instructions`
191
+ - Similar to the `lmsys-chat-1m-synth-ja-wo-pii-and-template-instructions`, but this version uses the original English user instructions. The assistant responses were generated in English as well. Rejection sampling was not applied in this version.
192
+ - filtered-magpie-ultra-en
193
+ - Subset of the [magpie-ultra](https://huggingface.co/datasets/argilla/magpie-ultra-v0.1) dataset, containing samples rated as 'average,' 'good,' or 'excellent.'.
194
+
195
  ## Risks and Limitations
196
 
197
  The models released here are still in the early stages of our research and development and have not been tuned to ensure outputs align with human intent and safety considerations.