File size: 1,372 Bytes
f84ad3c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
557f4b3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
---
dataset_info:
  features:
  - name: input_ids
    sequence: int32
  splits:
  - name: train
    num_bytes: 34065019260.0
    num_examples: 1039455
  download_size: 15973004443
  dataset_size: 34065019260.0
configs:
- config_name: default
  data_files:
  - split: train
    path: data/train-*
---

# OpenWebTextCorpus tokenized for Llama 3

This dataset is a pre-tokenized version of the [Skylion007/openwebtext](https://huggingface.co/datasets/Skylion007/openwebtext) dataset
using the [llama3](https://huggingface.co/meta-llama/Meta-Llama-3-8B) tokenizer. As such, this dataset follows the same licensing as the original openwebtext dataset.

This pre-tokenization is done as a performance optimization for using the openwebtext dataset with a Llama3 model.
This dataset was created using [SAELens](https://github.com/jbloomAus/SAELens), with the following settings:

- context_size: 8192
- shuffled: true
- begin_batch_token: "bos"
- begin_sequence_token: null
- sequence_separator_token: "eos"
- sae_lens_version: "3.3.0"

The `eos` token was used as a separator between sequences, since this resulted in the lowest loss experimentally.
Ideally we would like to use the same tokenization settings as used by the original Llama3 training regime, so if
you have information that the original Llama3 was trained using a different tokenization setup, please reach out!