Lemmy — Gemma 4 26B A4B MoE — MLX 4-bit

The Mixture-of-Experts member of the Lemma model family by Lethean. An EUPL-1.2 fork of Gemma 4 26B A4B with the Lethean Ethical Kernel (LEK) merged into the weights.

This repo hosts the MLX 4-bit build for native Apple Silicon inference via mlx-lm and mlx-vlm. For the GGUF playground (Ollama, llama.cpp) see lthn/lemmy. For the unmodified Google base see LetheanNetwork/lemmy.

Family

Repo Format Bits
lthn/lemmy GGUF multi-quant Q4_K_M → BF16
lthn/lemmy-mlx MLX 4-bit
lthn/lemmy-mlx-8bit MLX 8-bit
lthn/lemmy-mlx-bf16 MLX bf16

License

EUPL-1.2. See Gemma Terms of Use for upstream base model terms.

Downloads last month
309
Safetensors
Model size
25B params
Tensor type
BF16
·
U32
·
MLX
Hardware compatibility
Log In to add your hardware

Quantized

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for lthn/lemmy-mlx

Finetuned
lthn/lemmy
Quantized
(3)
this model

Collection including lthn/lemmy-mlx