Google DeepMind Publishes Gemma 3 Technical Report Detailing 27B Open Model
Google DeepMind releases the technical report for Gemma 3, its 27-billion-parameter open model that supports vision, text, and 140+ languages — with benchmark scores competitive with models significantly larger in size.
Google DeepMind has released the technical report for Gemma 3, its latest open model that features 27 billion parameters with support for vision, text generation, and over 140 languages. The report details the model's architecture, training methodology, and benchmark performance, which is competitive with significantly larger open-source models.
Architecture
Gemma 3 uses a dense transformer architecture with 27 billion parameters — a deliberate choice to remain at a size that can run on consumer-grade GPUs with 24GB of VRAM. The model supports both text-only and vision-language tasks through a unified architecture that processes images and text within the same model, rather than using separate vision encoders. This approach simplifies deployment and reduces the overhead of running multimodal AI applications.
Performance
On standard benchmarks, Gemma 3 performs competitively with models that are 2-3x larger, including Meta's Llama 3.3 70B and Alibaba's Qwen 3.5 72B. The model achieves particularly strong results on multilingual benchmarks, reflecting Google's investment in training data diversity across its 140+ supported languages. On coding benchmarks, Gemma 3 scores within a few percentage points of the best open-source models, making it a viable option for code-assisted development workflows.
Open Model Ecosystem
Gemma 3 is available under Google's permissive Gemma license, which allows commercial use, fine-tuning, and redistribution. The release continues Google's strategy of maintaining a competitive position in the open-source AI ecosystem alongside Meta's Llama, Alibaba's Qwen, and Mistral's models. For developers, the practical significance is that a 27B model that runs on a single consumer GPU can now perform tasks that required much larger models — or expensive API calls — just months ago, further reducing the cost of building AI-powered applications.
Related Articles
NVIDIA GTC 2026 Keynote: Jensen Huang Unveils Vera Rubin Platform and Six New Chips
NVIDIA CEO Jensen Huang opened GTC 2026 in San Jose with the formal unveiling of the complete Vera Rubin GPU platform — six new chips featuring 288 GB of HBM4 memory, 336 billion transistors, and 50 PetaFLOPS of FP4 performance. Over 30,000 attendees from 190 countries gathered for the AI industry's most anticipated annual event.
OpenAI Acquires Promptfoo to Strengthen AI Agent Security and Red-Teaming
OpenAI has agreed to acquire Promptfoo, the open-source AI security and red-teaming platform used by over 25% of the Fortune 500, in a deal that will integrate the tool directly into OpenAI's enterprise agent platform. The acquisition signals OpenAI's growing focus on safety infrastructure as it pushes deeper into autonomous AI agent deployment.
NVIDIA Releases Nemotron 3 Super: Open 120B-Parameter Model Targets Enterprise Agentic AI
NVIDIA has released Nemotron 3 Super, a 120-billion-parameter open-weights model built on a hybrid Mamba-Transformer architecture with a one-million-token context window. The model delivers 5x throughput improvements over its predecessor and is designed specifically for enterprise agentic AI workflows.