Hugging Face has made a significant advancement in the field of small language models (SLMs) with the release of SmolLM2, a compact 1.7 billion parameter language model. SmolLM2, which sets a new benchmark for performance among models of its size, is designed for on-device use, handling texts up to 20 pages, generating concise summaries, and supporting interactive follow-up questions. In addition, Hugging Face introduced SmolTools, a suite of lightweight AI-powered tools built with LLaMA.cpp and small language models, showcasing the capabilities of these new models. SmolLM2 is available in three sizes: 135M, 360M, and 1.7B. Other companies are also accelerating their SLM efforts, with Meta open-sourcing the MobileLLM foundation model for on-device scenarios. The rise of small language models is driven by their efficiency and ability to perform tasks typically handled by larger models, making them an attractive option for businesses aiming to operationalize responsible AI.
A recurring theme we hear from businesses looking to operationalize responsible AI is this: How can we create the smallest models possible that are both effective and explainable? Large language models (LLMs) are essentially black boxes. They might be useful for general tasks…
A Comprehensive Survey of Small Language Models Nice survey on small language models (SLMs) and discussion on issues related to definitions, applications, enhancements, reliability, and more. https://t.co/qVxuY1jWDE https://t.co/WZuRm1fqU4
SMOLLM2, with sizes as small as 0.1B parameters, is reshaping AI. Explore the innovation behind its efficiency. Here’s how it outperforms larger models: Introducing SMOLLM2 @huggingface unveils SMOLLM2, a new lineup of small language models (SLMs) optimized for on-device… https://t.co/4ORL0Nn8dA