It's a super tiny llama3 model.

It has 0.247B parameters.

It is pretrained on the fineweb-edu dataset.(10B)

I hope I can make it beter and better.

If you see it, please give me a like. Thanks.

More info will be added later.

Downloads last month
17
Safetensors
Model size
313M params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Amu/supertiny-llama3-0.25B-v0.1

Quantizations
1 model

Dataset used to train Amu/supertiny-llama3-0.25B-v0.1

Collection including Amu/supertiny-llama3-0.25B-v0.1