It's a super tiny llama3 model.

It has 0.247B parameters.

It is pretrained on the fineweb-edu dataset.(10B)

I hope I can make it beter and better.

If you see it, please give me a like. Thanks.

More info will be added later.

Downloads last month
154
Safetensors
Model size
313M params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for Amu/supertiny-llama3-0.25B-v0.1

Quantizations
1 model

Dataset used to train Amu/supertiny-llama3-0.25B-v0.1

Collection including Amu/supertiny-llama3-0.25B-v0.1