Google Launches Open Model Family Gemma 4
The new offering is designed for advanced reasoning and multimodal capabilities.
Google launched Gemma 4, a new family of open-weight AI models that the company says are its “most intelligent” models to date.
The April 2 release builds on growing adoption of the Gemma series, which has seen more than 400 million downloads and 100,000 community-built variants since its debut in February 2024.
Designed to deliver major advancements in reasoning, code generation and complex logic tasks, Google says the family is built using the same research and tech as Gemini 3 but marks a significant leap forward in capabilities.
Gemma 4 is available in four sizes to meet different environmental criteria. The smaller 2-billion- and 4-billion-parameter "Effective" models are intended for edge devices such as smartphones, while the 26-billion-parameter mixture-of-experts and 31-billion-parameter dense models can be deployed in more compute-intensive workloads.
Parameters are the settings a large language model can use to generate an output. While more parameters yield better results, they also require more compute to run. With Gemma 4, Google said it hopes to offer a more effective alternative to previous models, achieving "an unprecedented level of intelligence-per-parameter."
Related:Microsoft Goes Beyond LLMs With New Voice, Image Models
“For developers, this new level of intelligence-per-parameter means achieving frontier-level capabilities with significantly less hardware overhead,” the vendor said in a blog post.
The models have been trained on more than 140 languages and feature built-in audio and visual processing, enabling them to run offline.
Despite their relatively modest size, Google claims the larger models deliver near-frontier performance, with the 26B variant outperforming models up to 20 times its size.
Accessibility is also a key differentiator with this release, with Gemma 4 distributed under an Apache 2.0 license (previous iterations were available through Google’s Gemma license). The change gives developers more freedom to modify and deploy the models commercially.
“This open source license provides a foundation for complete developer flexibility and digital sovereignty -- granting you complete control over your data, infrastructure, and models,” Google said. “It allows you to build freely and deploy securely across any environment, whether on premises or in the cloud.”
About the Author
Contributing Writer
Scarlett Evans is a freelance writer with a focus on emerging technologies and the minerals industry. Previously, she served as assistant editor at IoT World Today, where she specialized in robotics and smart city technologies. Scarlett also has a background in the mining and resources sector, with experience at Mine Australia, Mine Technology and Power Technology. She joined Informa in April 2022 before transitioning to freelance work.
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
modellaunchreasoning
AI Products Have Terrible UX: Here's Why
If you're building an AI product right now, the single most high-leverage thing you can do isn't to upgrade your model. It's to watch five users try to use your product without any help from you. Don't say anything. Just watch. You'll see exactly where the AI ends, and the confusion begins. That's your design debt. Read All

Launch HN: Freestyle – Sandboxes for Coding Agents
We’re Ben and Jacob, cofounders of Freestyle ( https://freestyle.sh ). We’re building a cloud for Coding Agents. For the first generation of agents it looked like workflows with minimal tools. 2 years ago we published a package to let AI work in SQL, at that time GPT-4 could write simple scripts. Soon after the first AI App Builders started using AI to make whole websites; we supported that with a serverless deploy system. But the current generation is going much further, instead of minimal tools and basic serverless apps AI can utilize the full power of a computer (“sandbox”). We’re building sandboxes that are interchangeable with EC2s from your agents perspective, with bonus features: 1. We’ve figured out how to fork a sandbox horizontally without more than a 400ms pause in it. That's no
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Frontier Research

Nirvana Rank (working post)
not meant to be formal/rigorous - more stream of thought, I'm trying to think about anthropics more I wanted to communicate a concept I'd like to call **Nirvana Rank** that I came up with when thinking about Anthropics and reference classes. I'll cover some background Anthropics is about reasoning from the fact that you're an observer of reality. It treats your own existence as evidence you can use to deduce things about the worlds you're in. Within anthropics, people have clustered on two ways of treating your own existence as evidence: The first is called the Self Sampling Assumption (SSA) - where you treat yourself as a random sample, a typical member from the space of actual observers that exist in your reference class. The second is called the Self Indication Assumption (SIA) which co





Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!