Google's Gemma 4 Brings 256K Context to Open Models

April 4, 2026

Google's Gemma 4 Brings 256K Context to Open Models

Published: April 4, 2026 at 12:37 AM

Updated: April 4, 2026 at 12:37 AM

100-word summary

Google just released Gemma 4, a four-model family under the Apache 2.0 license. The headline feature: context windows up to 256K tokens, letting open models chew through entire codebases or legal documents without proprietary API bills. The smallest model handles 128K contexts on edge hardware. It's available now on Hugging Face, Kaggle, and Ollama with native function-calling for building agents that interact with tools. The 26B mixture-of-experts model activates only 3.8B parameters per query, keeping responses fast. What's striking is the pairing: truly open licensing with capabilities previously locked behind commercial models. Open-source AI just got a lot more competitive with closed alternatives.

What happened

Google just released Gemma 4, a four-model family under the Apache 2.0 license. The headline feature: context windows up to 256K tokens, letting open models chew through entire codebases or legal documents without proprietary API bills. The smallest model handles 128K contexts on edge hardware. It's available now on Hugging Face, Kaggle, and Ollama with native function-calling for building agents that interact with tools. The 26B mixture-of-experts model activates only 3.8B parameters per query, keeping responses fast. What's striking is the pairing: truly open licensing with capabilities previously locked behind commercial models.

Why it matters

Open-source AI just got a lot more competitive with closed alternatives.

Sources