Update README.md
Browse files
README.md
CHANGED
@@ -6,9 +6,12 @@ library_name: transformers
|
|
6 |
pipeline_tag: text-generation
|
7 |
license: llama3.1
|
8 |
model_type: llama
|
|
|
|
|
|
|
9 |
---
|
10 |
|
11 |
-
#
|
12 |
|
13 |
Llama 3.1 Swallow is a series of large language models (8B, 70B) that were built by continual pre-training on the [Meta Llama 3.1](https://huggingface.co/collections/meta-llama/llama-31-669fc079a0c406a149a5738f) models.
|
14 |
Llama 3.1 Swallow enhanced the Japanese language capabilities of the original Llama 3.1 while retaining the English language capabilities.
|
@@ -197,20 +200,23 @@ print(output[0].outputs[0].text)
|
|
197 |
|
198 |
### Instruction Tuning
|
199 |
|
200 |
-
The following datasets were used for the instruction tuning.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
201 |
|
202 |
-
- lmsys-chat-1m-synth-ja-wo-pii
|
203 |
-
|
204 |
-
- Japanese translation of the lmsys-chat-1m dataset using DeepL, with synthetic instruction data created using the Llama-3.1-405B model.
|
205 |
-
- 'wo-pii' indicates removal of personally identifiable information.
|
206 |
|
207 |
-
- filtered magpie-ultra
|
208 |
-
- Subset of the [magpie-ultra](https://huggingface.co/datasets/argilla/magpie-ultra-v0.1) dataset, containing samples rated as 'average,' 'good,' or 'excellent.'.
|
209 |
-
|
210 |
-
- gemma-magpie
|
211 |
-
- Japanese dataset.
|
212 |
-
- Generated using prompts for specific category words.
|
213 |
-
|
214 |
## Risks and Limitations
|
215 |
|
216 |
The models released here are still in the early stages of our research and development and have not been tuned to ensure outputs align with human intent and safety considerations.
|
|
|
6 |
pipeline_tag: text-generation
|
7 |
license: llama3.1
|
8 |
model_type: llama
|
9 |
+
datasets:
|
10 |
+
- lmsys/lmsys-chat-1m
|
11 |
+
- argilla/magpie-ultra-v0.1
|
12 |
---
|
13 |
|
14 |
+
# Llama 3.1 Swallow
|
15 |
|
16 |
Llama 3.1 Swallow is a series of large language models (8B, 70B) that were built by continual pre-training on the [Meta Llama 3.1](https://huggingface.co/collections/meta-llama/llama-31-669fc079a0c406a149a5738f) models.
|
17 |
Llama 3.1 Swallow enhanced the Japanese language capabilities of the original Llama 3.1 while retaining the English language capabilities.
|
|
|
200 |
|
201 |
### Instruction Tuning
|
202 |
|
203 |
+
The following instruction datasets were used for the instruction tuning.
|
204 |
+
- Japanese
|
205 |
+
- `lmsys-chat-1m-synth-ja-wo-pii-and-template-instructions`
|
206 |
+
- Single-turn Japanese synthetic instruction dataset derived from [lmsys-chat-1m](https://huggingface.co/datasets/lmsys/lmsys-chat-1m) dataset [\[Zhang+, ICLR24\]](https://openreview.net/forum?id=BOfDKxfwt0)).
|
207 |
+
The first-turn user instructions were translated into Japanese via DeepL machine translation, and the assistant responses were generated using the Llama 3.1 405B Instruct model. Rejection sampling (n=6) was applied, with Llama 3.1 70B Instruct serving as a judge.
|
208 |
+
- As implied by the dataset name, conversations that contain personally identifiable information (PII) or template-based user instructions have been removed. Duplicate instuctions have also been removed.
|
209 |
+
- `filtered-magpie-ultra-ja`
|
210 |
+
- A Japanese variant of the `filtered-magpie-ultra-en` dataset, machine-translated into Japanese using the Gemma 2 27B IT.
|
211 |
+
- `gemma-magpie`
|
212 |
+
- Japanese Q&A dataset on diverse topics, generated using prompts with specific category words, with answers by Gemma 2 27B IT, heuristically filtered for quality and length.
|
213 |
+
- English
|
214 |
+
- `lmsys-chat-1m-synth-en-wo-pii-and-template-instructions`
|
215 |
+
- Similar to the `lmsys-chat-1m-synth-ja-wo-pii-and-template-instructions`, but this version uses the original English user instructions. The assistant responses were generated in English as well. Rejection sampling was not applied in this version.
|
216 |
+
- `filtered-magpie-ultra-en`
|
217 |
+
- A subset of the [magpie-ultra](https://huggingface.co/datasets/argilla/magpie-ultra-v0.1) dataset, developed following the MAGPIE recipe [\[Xu+, arXiv24\]](https://arxiv.org/abs/2406.08464) using Llama 3.1 405B Instruct. This subset includes only samples rated as 'average,' 'good,' or 'excellent.'
|
218 |
|
|
|
|
|
|
|
|
|
219 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
220 |
## Risks and Limitations
|
221 |
|
222 |
The models released here are still in the early stages of our research and development and have not been tuned to ensure outputs align with human intent and safety considerations.
|