Linked to instruction datasets, updted metadata, linked to 70B Instruct v0.3.
Browse files
README.md
CHANGED
@@ -12,6 +12,9 @@ datasets:
|
|
12 |
- lmsys/lmsys-chat-1m
|
13 |
- tokyotech-llm/lmsys-chat-1m-synth
|
14 |
- argilla/magpie-ultra-v0.1
|
|
|
|
|
|
|
15 |
---
|
16 |
|
17 |
# Llama 3.1 Swallow - Built with Llama
|
@@ -36,7 +39,7 @@ See the Swallow Model Index section to find other model variants.
|
|
36 |
|Model|Llama-3.1-Swallow v0.1|Llama-3.1-Swallow-Instruct v0.1|Llama-3.1-Swallow v0.2|Llama-3.1-Swallow-Instruct v0.2|Llama-3.1-Swallow-Instruct v0.3|
|
37 |
|---|---|---|---|---|---|
|
38 |
|8B| [Link](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-v0.1) | [Link](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.1) | [Link](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-v0.2) | [Link](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.2) | [Link](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3)
|
39 |
-
|70B| [Link](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-70B-v0.1) | [Link](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-70B-Instruct-v0.1) | | | |
|
40 |
|
41 |

|
42 |
|
@@ -196,22 +199,19 @@ print(output[0].outputs[0].text)
|
|
196 |
The following datasets were used for the instruction tuning.
|
197 |
|
198 |
- Japanese
|
199 |
-
-
|
200 |
- Single-turn Japanese instruction dataset synthesized and derived from [lmsys-chat-1m](https://huggingface.co/datasets/lmsys/lmsys-chat-1m) [\[Zhang+, ICLR24\]](https://openreview.net/forum?id=BOfDKxfwt0)). First-turn user instructions were translated into Japanese via DeepL (machine translation), and assistant responses were generated using [Llama-3.1-405B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-405B-Instruct). [Llama-3.1-70B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-70B-Instruct) served as a judge for rejection sampling (n=6).
|
201 |
Conversations containing personally identifiable information (PII) and template-based user instructions were removed. Duplicate instructions were removed.
|
202 |
-
|
203 |
-
- `filtered-magpie-ultra-ja`
|
204 |
- A Japanese variant of the `filtered-magpie-ultra-en` dataset, translated into Japanese by [gemma-2-27b-it](https://huggingface.co/google/gemma-2-27b-it).
|
205 |
-
-
|
206 |
-
- A Japanese synthetic
|
207 |
- English
|
208 |
-
-
|
209 |
-
- The creation process is similar to `
|
210 |
-
- The dataset is available at [tokyotech-llm/lmsys-chat-1m-synth](https://huggingface.co/datasets/tokyotech-llm/lmsys-chat-1m-synth).
|
211 |
- `filtered-magpie-ultra-en`
|
212 |
- A subset of the [magpie-ultra](https://huggingface.co/datasets/argilla/magpie-ultra-v0.1) dataset, developed following the MAGPIE recipe [\[Xu+, arXiv24\]](https://arxiv.org/abs/2406.08464) using [Llama-3.1-405B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-405B-Instruct). This subset includes only samples rated as 'average,' 'good,' or 'excellent.'
|
213 |
|
214 |
-
|
215 |
## Risks and Limitations
|
216 |
|
217 |
The models released here are still in the early stages of our research and development and have not been tuned to ensure outputs align with human intent and safety considerations.
|
|
|
12 |
- lmsys/lmsys-chat-1m
|
13 |
- tokyotech-llm/lmsys-chat-1m-synth
|
14 |
- argilla/magpie-ultra-v0.1
|
15 |
+
- tokyotech-llm/swallow-magpie-ultra-v0.1
|
16 |
+
- tokyotech-llm/swallow-gemma-magpie-v0.1
|
17 |
+
|
18 |
---
|
19 |
|
20 |
# Llama 3.1 Swallow - Built with Llama
|
|
|
39 |
|Model|Llama-3.1-Swallow v0.1|Llama-3.1-Swallow-Instruct v0.1|Llama-3.1-Swallow v0.2|Llama-3.1-Swallow-Instruct v0.2|Llama-3.1-Swallow-Instruct v0.3|
|
40 |
|---|---|---|---|---|---|
|
41 |
|8B| [Link](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-v0.1) | [Link](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.1) | [Link](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-v0.2) | [Link](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.2) | [Link](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3)
|
42 |
+
|70B| [Link](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-70B-v0.1) | [Link](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-70B-Instruct-v0.1) | | | [Link](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-70B-Instruct-v0.3) |
|
43 |
|
44 |

|
45 |
|
|
|
199 |
The following datasets were used for the instruction tuning.
|
200 |
|
201 |
- Japanese
|
202 |
+
- [Llama-3.1-LMSYS-Chat-1M-Synth-Ja](https://huggingface.co/datasets/tokyotech-llm/lmsys-chat-1m-synth)
|
203 |
- Single-turn Japanese instruction dataset synthesized and derived from [lmsys-chat-1m](https://huggingface.co/datasets/lmsys/lmsys-chat-1m) [\[Zhang+, ICLR24\]](https://openreview.net/forum?id=BOfDKxfwt0)). First-turn user instructions were translated into Japanese via DeepL (machine translation), and assistant responses were generated using [Llama-3.1-405B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-405B-Instruct). [Llama-3.1-70B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-70B-Instruct) served as a judge for rejection sampling (n=6).
|
204 |
Conversations containing personally identifiable information (PII) and template-based user instructions were removed. Duplicate instructions were removed.
|
205 |
+
- [Swallow-Magpie-Ultra-v0.1](https://huggingface.co/datasets/tokyotech-llm/swallow-magpie-ultra-v0.1)
|
|
|
206 |
- A Japanese variant of the `filtered-magpie-ultra-en` dataset, translated into Japanese by [gemma-2-27b-it](https://huggingface.co/google/gemma-2-27b-it).
|
207 |
+
- [Swallow-Gemma-Magpie-v0.1](https://huggingface.co/datasets/tokyotech-llm/swallow-gemma-magpie-v0.1)
|
208 |
+
- A Japanese synthetic instruction tuning dataset from scratch, generated by [gemma-2-27b-it](https://huggingface.co/google/gemma-2-27b-it). User instructions were created with prompts specific to each topic, and assistant responses were generated for these instructions. The conversations were then heuristically filtered for quality and length.
|
209 |
- English
|
210 |
+
- [Llama-3.1-LMSYS-Chat-1M-Synth-En](https://huggingface.co/datasets/tokyotech-llm/lmsys-chat-1m-synth)
|
211 |
+
- The creation process is similar to `Llama-3.1-LMSYS-Chat-1M-Synth-Ja`, but this version uses the original English user instructions. The assistant responses were generated in English as well. Rejection sampling was not applied for this version.
|
|
|
212 |
- `filtered-magpie-ultra-en`
|
213 |
- A subset of the [magpie-ultra](https://huggingface.co/datasets/argilla/magpie-ultra-v0.1) dataset, developed following the MAGPIE recipe [\[Xu+, arXiv24\]](https://arxiv.org/abs/2406.08464) using [Llama-3.1-405B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-405B-Instruct). This subset includes only samples rated as 'average,' 'good,' or 'excellent.'
|
214 |
|
|
|
215 |
## Risks and Limitations
|
216 |
|
217 |
The models released here are still in the early stages of our research and development and have not been tuned to ensure outputs align with human intent and safety considerations.
|