nielsr's picture
nielsr HF staff
Add model card metadata: pipeline tag, library name, link to paper, and link to code repository.
9db2f00 verified
|
raw
history blame
1.47 kB
metadata
license: apache-2.0
library_name: transformers
pipeline_tag: text-generation

From scratch pretraining on english only no synthetic data, no code, 3 epochs of 1 gig of data for the ~125M param model.

Test network using Tensor Product Attention. Other than some alterations to the attention, such as 16 heads insted of 9 and using TPA, this is the same setup as https://huggingface.co/HuggingFaceTB/SmolLM2-135M-Instruct

Scripts:

  • inference.py to run the model with some test prompts
  • test_train.py runs with the exact configurations used to train this model and is the reproduction script. Data is assumed to be in JSONL format with "text":"example text", "text":"..."

Notes:

One of the primary reported benefits for TPA are for inference which are not really being leveraged at all, although you can probably fit a larger bsz than traditional MHA/GQA with this. This did save about 5% on params, that amount should scale much more as the network size increases. The run time is very similar to MHA/GQA at this scale.

Training Metrics

Dataset Information

  • Training data per epoch: 1 GB
  • Total tokens trained: 48,261,120
  • No sythetic data

Training Results

  • Final Train Loss: 3.0421
  • Final Train Perplexity: 20.95

image/png

Code

The code is available at: https://github.com/tensorgi/T6.