Google’s DeepMind research arm on 14 August unveiled Gemma 3 270M, an open-weight language model that packs 270 million parameters into a download of roughly 240 MB. The model is the smallest member of the Gemma 3 family and is aimed at developers who need fast, low-cost inference on laptops, edge devices and smartphones. Built on the Gemma 3 architecture, the system allocates 170 million parameters to embeddings and 100 million to transformer blocks, backed by a 256 k-token vocabulary to handle rare terms. An instruction-tuned variant scored about 51 percent on the IFEval benchmark, outperforming several models with similar or larger footprints, according to figures released by Google. Energy consumption is a principal selling point: internal tests show an INT4-quantised version consuming just 0.75 percent of a Pixel 9 Pro’s battery across 25 conversational sessions. The company says Gemma 3 270M supports rapid fine-tuning in minutes and can run fully offline in browsers, Raspberry Pi boards and other low-power hardware. Google is releasing both pre-trained and instruction-tuned checkpoints, plus Quantization-Aware Training versions, under its Gemma licence that permits commercial use subject to safety restrictions. Documentation and deployment recipes for tools such as Hugging Face, Ollama and JAX are provided to accelerate adoption by enterprises and independent developers.
Run Google's Gemma-3-270M in Jan - Click the Use this model button: https://t.co/n7ghwpuTOi - Select 👋 Jan - Download & start using the model Shoutout @ggml_org for the GGUF 💙 https://t.co/oDKSjGCeng
Gemma 3 270M running on my Pixel 7a! Absolutely crazy (not sped up) https://t.co/zhYvNmCdjA https://t.co/3Fda5vww9R
Google just dropped a new tiny LLM with outstanding performance -- Gemma3 270M. Now available on KerasHub. Try the new presets `gemma3_270m` and `gemma3_instruct_270m`! https://t.co/1ttd41rpkS