This is a randomly created qwen3 tiny model with 10M parameters. This is part of a projct to create a LLM from scratch completely on apple silicon. I'm using the created tokenizer from alibaba, thats used for the Qwen3 models.

The next (trained) versino of this will be on Goekdeniz-Guelmez/J.O.S.I.E.-Qwen3-10M-Base-Phase1.

Downloads last month
8
Safetensors
Model size
10M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Goekdeniz-Guelmez/J.O.S.I.E.-Qwen3-10M-Random

Finetunes
1 model