Spaces:
Running
Running
<html lang="en"> | |
<head> | |
<meta charset="UTF-8"> | |
<meta name="viewport" content="width=device-width, initial-scale=1"> | |
<title>Top Open-Source Small Language Models</title> | |
<link rel="stylesheet" href="styles.css"/> | |
</head> | |
<body> | |
<h1>Top Open-Source Small Language Models for Generative AI Applications</h1> | |
<p> | |
Small Language Models (SLMs) are language models that contain, at most, a few billion parametersβsignificantly fewer | |
than Large Language Models (LLMs), which can have tens, hundreds of billions, or even trillions, of parameters. SLMs | |
are well-suited for resource-constrained environments, as well as on-device and real-time generative AI | |
applications. Many of them can run locally on a laptop using tools like LM Studio or Ollama . These models are | |
typically derived from larger models using techniques such as quantization and distillation. In the following, some | |
well developed SLMs are introduced. | |
</p> | |
<p> | |
Note: All the models mentioned here are open source. However, for details regarding experimental use, commercial | |
use, redistribution, and other terms, please refer to the license documentation. | |
</p> | |
<h2>Phi 4 Collection by Microsoft</h2> | |
<p> | |
This Collection features a range of small language models, including reasoning models, ONNX- and GGUF-compatible | |
formats, and multimodal models. The base model in the collection has 14 billion parameters, while the smallest | |
models have 3.84 billion. Strategic use of synthetic data during training has led to improved performance compared | |
to its mother model (primarily GPT-4). Currently, the collection includes three versions of reasoning-focused SLMs, | |
making it one of the best solutions for reasoning tasks. | |
</p> | |
<p> | |
π Licence: <a href="https://choosealicense.com/licenses/mit/" target="_blank">MIT</a><br> | |
π <a href="https://huggingface.co/collections/microsoft/phi-4-677e9380e514feb5577a40e4" target="_blank">Collection | |
on Hugging Face</a><br> | |
π <a href="https://arxiv.org/abs/2412.08905" target="_blank">Technical Report</a> | |
</p> | |
<h2>Gemma 3 Collection by Google</h2> | |
<p> | |
This collection features multiple versions, including Image-to-Text, Text-to-Text, and Image-and-Text-to-Text | |
models, available in both quantized and GGUF formats. The models vary in size, with 1, 4.3, 12.2, and 27.4 billion | |
parameters. Two specialized variants have been developed for specific applications: TxGemma, optimized for | |
therapeutic development, and ShieldGemma, designed for moderating text and image content. | |
</p> | |
<p> | |
π Licence: <a href="https://ai.google.dev/gemma/terms" target="_blank">Gemma</a><br> | |
π <a href="https://huggingface.co/collections/google/gemma-3-release-67c6c6f89c4f76621268bb6d" target="_blank">Collection | |
on Hugging Face</a><br> | |
π <a href="https://storage.googleapis.com/deepmind-media/gemma/Gemma3Report.pdf" target="_blank">Technical | |
Report</a><br> | |
π <a href="https://huggingface.co/collections/google/shieldgemma-67d130ef8da6af884072a789" target="_blank">ShieldGemma | |
on Hugging Face</a><br> | |
π <a href="https://huggingface.co/collections/google/txgemma-release-67dd92e931c857d15e4d1e87" target="_blank">TxGemma | |
on Hugging Face</a> | |
</p> | |
<h2>Mistral Models</h2> | |
<p> | |
Mistral AI is a France-based AI startup and one of the pioneers in releasing open-source language models. Its | |
current product lineup includes three compact models: Mistral Small 3.1, Pixtral 12B, and Mistral NEMO. All of them | |
are released under <a href="https://www.apache.org/licenses/LICENSE-2.0" target="_blank">Apache 2.0 license</a>. | |
</p> | |
<p> | |
<b>Mistral 3.1</b> is a multimodal and multilingual SLM having 24 billion parameters and 128K context window. | |
Currently there are two versions: Base and Instruct.<br> | |
π <a href="https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Base-2503" target="_blank">Base Version on Hugging | |
Face</a><br> | |
π <a href="https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Instruct-2503" target="_blank">Instruct Version on | |
Hugging Face</a><br> | |
π <a href="https://mistral.ai/news/mistral-small-3-1" target="_blank">Technical Report</a><br> | |
</p> | |
<p> | |
<b>Pixtral 12B</b> is a natively multimodal model trained on interleaved image and text data, delivering strong | |
performance on multimodal tasks and instruction following while maintaining state-of-the-art results on text-only | |
benchmarks. It features a newly developed 400M parameter vision encoder and a 12B parameter multimodal decoder based | |
on Mistral NEMO. The model supports variable image sizes, aspect ratios, and multiple images within a long context | |
window of up to 128k tokens.<br> | |
π <a href="https://huggingface.co/mistralai/Pixtral-12B-Base-2409" target="_blank">Pixtral-12B-Base-2409 on Hugging | |
Face</a><br> | |
π <a href="https://huggingface.co/mistralai/Pixtral-12B-2409" target="_blank">Pixtral-12B-2409 on Hugging | |
Face</a><br> | |
π <a href="https://mistral.ai/news/pixtral-12b" target="_blank">Technical Report</a><br> | |
</p> | |
<p> | |
<b>Mistral NeMo</b> is a 12B model developed in collaboration with NVIDIA, featuring a large 128k-token context | |
window and state-of-the-art reasoning, knowledge, and coding accuracy for its size.<br> | |
π <a href="https://huggingface.co/mistralai/Mistral-Nemo-Instruct-FP8-2407" target="_blank">Model on Hugging | |
Face</a><br> | |
π <a href="https://mistral.ai/news/mistral-nemo" target="_blank">Technical Report</a> | |
</p> | |
<h2>Llama Models by Meta</h2> | |
<p> | |
Meta is one of the leading contributors to open-source AI. In recent years, it has released several versions of its | |
Llama models. The latest series is Llama 4, although all models in this collection are currently quite large. | |
Smaller models may be introduced in the future or in upcoming sub-versions of Llama 4, but for now, that hasnβt | |
happened. The most recent collection that includes smaller models is Llama 3.2. It features models with 1.24 billion | |
and 3.21 billion parameters with 128k context windows. Additionally, there is a 10.6 billion-parameter multimodal | |
version designed for Image-and-Text-to-Text tasks. | |
This collection includes small variants of Llama Guard β fine-tuned language models designed for prompt and response | |
classification. They can detect unsafe prompts and responses, making them useful for implementing safety measures in | |
LLM-based applications. | |
</p> | |
<p> | |
π License: <a href="https://www.llama.com/llama3_2/license/" target="_blank">LLAMA 3.2 COMMUNITY LICENSE | |
AGREEMENT</a><br> | |
π <a href="https://huggingface.co/collections/meta-llama/llama-32-66f448ffc8c32f949b04c8cf" target="_blank">Collection | |
on Hugging Face</a><br> | |
π <a href="https://ai.meta.com/blog/llama-3-2-connect-2024-vision-edge-mobile-devices/" target="_blank">Technical | |
Paper</a> | |
</p> | |
<h2>Qwen 3 Collection by Alibaba</h2> | |
<p> | |
The Chinese tech giant Alibaba is another major player in open-source AI. It releases its language models under the | |
Qwen name. The latest version is Qwen 3, which includes both small and large models. The smaller models range in | |
size, with parameter counts of 14.8 billion, 8.19 billion, 4.02 billion, 2.03 billion, and even 752 million. This | |
collection also includes quantized and GGUF formats. | |
</p> | |
<p> | |
π Licence: <a href="https://www.apache.org/licenses/LICENSE-2.0" target="_blank">Apache 2.0</a><br> | |
π <a href="https://huggingface.co/collections/Qwen/qwen3-67dd247413f0e2e4f653967f" target="_blank">Collection on | |
Hugging Face</a><br> | |
π <a href="https://github.com/QwenLM/Qwen3/blob/main/Qwen3_Technical_Report.pdf" target="_blank">Technical | |
Report</a> | |
</p> | |
<hr style="border: none; height: 1px; background-color: #ccc;"> | |
<p>This list is not limited to these five. You can explore more open-source models at:</p> | |
<ul> | |
<li><a href="https://huggingface.co/databricks" target="_blank">Databricks</a></li> | |
<li><a href="https://huggingface.co/Cohere" target="_blank">Cohere</a></li> | |
<li><a href="https://huggingface.co/deepseek-ai" target="_blank">Deepseek</a></li> | |
<li><a href="https://huggingface.co/collections/HuggingFaceTB/smollm-6695016cad7167254ce15966" target="_blank">SmolLM</a> | |
</li> | |
<li><a href="https://huggingface.co/stabilityai" target="_blank">Stability AI</a></li> | |
<li><a href="https://huggingface.co/ibm-granite" target="_blank">IBM Granite</a></li> | |
</ul> | |
</body> | |
</html> | |