|
--- |
|
dataset_info: |
|
features: |
|
- name: input_ids |
|
sequence: int32 |
|
splits: |
|
- name: train |
|
num_bytes: 34065019260.0 |
|
num_examples: 1039455 |
|
download_size: 15973004443 |
|
dataset_size: 34065019260.0 |
|
configs: |
|
- config_name: default |
|
data_files: |
|
- split: train |
|
path: data/train-* |
|
--- |
|
|
|
# OpenWebTextCorpus tokenized for Llama 3 |
|
|
|
This dataset is a pre-tokenized version of the [Skylion007/openwebtext](https://huggingface.co/datasets/Skylion007/openwebtext) dataset |
|
using the [llama3](https://huggingface.co/meta-llama/Meta-Llama-3-8B) tokenizer. As such, this dataset follows the same licensing as the original openwebtext dataset. |
|
|
|
This pre-tokenization is done as a performance optimization for using the openwebtext dataset with a Llama3 model. |
|
This dataset was created using [SAELens](https://github.com/jbloomAus/SAELens), with the following settings: |
|
|
|
- context_size: 8192 |
|
- shuffled: true |
|
- begin_batch_token: "bos" |
|
- begin_sequence_token: null |
|
- sequence_separator_token: "eos" |
|
- sae_lens_version: "3.3.0" |
|
|
|
The `eos` token was used as a separator between sequences, since this resulted in the lowest loss experimentally. |
|
Ideally we would like to use the same tokenization settings as used by the original Llama3 training regime, so if |
|
you have information that the original Llama3 was trained using a different tokenization setup, please reach out! |
|
|