{"id":382924,"date":"2026-04-16T18:53:17","date_gmt":"2026-04-16T18:53:17","guid":{"rendered":"https:\/\/www.newsbeep.com\/nz\/382924\/"},"modified":"2026-04-16T18:53:17","modified_gmt":"2026-04-16T18:53:17","slug":"google-opens-gemma-4-under-apache-2-0-with-multimodal-and-agentic-capabilities","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/nz\/382924\/","title":{"rendered":"Google Opens Gemma 4 Under Apache 2.0 with Multimodal and Agentic Capabilities"},"content":{"rendered":"<p>Google has recently <a href=\"https:\/\/blog.google\/innovation-and-ai\/technology\/developers-tools\/gemma-4\/\" rel=\"nofollow noopener\" target=\"_blank\">released<\/a> <a href=\"https:\/\/deepmind.google\/models\/gemma\/gemma-4\/\" rel=\"nofollow noopener\" target=\"_blank\">Gemma 4<\/a>, a family of open-weight models spanning effective 2B and 4B edge variants, a 26B Mixture-of-Experts model, and a 31B dense model, all distributed under an <a href=\"https:\/\/www.apache.org\/licenses\/LICENSE-2.0\" rel=\"nofollow noopener\" target=\"_blank\">Apache 2.0 license<\/a>. The release introduces native video and image processing across the lineup, audio input on the smaller models, context windows up to 256K tokens, and benchmark results that place the 31B dense variant in a bracket typically occupied by models three to five times its size.<\/p>\n<p>The agent-oriented framing is reflected in concrete capabilities. Google reports that the 31B variant scores <a href=\"https:\/\/ai.google.dev\/gemma\/docs\/core\/model_card_4\" rel=\"nofollow noopener\" target=\"_blank\">84.3% on GPQA Diamond and 80.0% on LiveCodeBench v6<\/a>. The GPQA Diamond result nearly doubles <a href=\"https:\/\/ai.google.dev\/gemma\/docs\/core\/model_card_3#benchmark_results\" rel=\"nofollow noopener\" target=\"_blank\">the 42.4% achieved by the prior Gemma 3 IT 27B<\/a>, reflecting substantial gains in science reasoning and code generation. For tool use, the models add native support for <a href=\"https:\/\/docs.cloud.google.com\/vertex-ai\/generative-ai\/docs\/multimodal\/function-calling\" rel=\"nofollow noopener\" target=\"_blank\">function-calling<\/a>, structured JSON output, and native system instructions, a combination intended to let developers build autonomous agents that interact with external tools and APIs and execute multi-step workflows reliably.<\/p>\n<p>Architecturally, the lineup spans both dense and sparse designs. The 26B MoE model activates only 3.8 billion parameters during inference to deliver fast tokens-per-second, while the 31B dense variant targets workloads where consistent per-token cost matters more than peak parameter count. The edge models, sized for mobile and IoT devices where memory and power budgets are tight, offer a 128K context window; the larger models extend to 256K tokens, large enough to ingest sizable code repositories or long-form documents in a single prompt. All four variants natively process video and images at variable resolutions, and the E2B and E4B edge models add native audio input for speech recognition and understanding, and the family is trained on more than 140 languages.<\/p>\n<p>On benchmarks, Google reports that the 31B dense model achieved an estimated <a href=\"https:\/\/arena.ai\/leaderboard\/text?license=open-source\" rel=\"nofollow noopener\" target=\"_blank\">LLMArena<\/a> score (text only) of 1452, reaching a performance bracket usually reserved for significantly larger models with triple to quintuple the parameter count.<\/p>\n<p><img decoding=\"async\" alt=\"\" src=\"https:\/\/www.infoq.com\/news\/2026\/04\/google-gemm4\/news\/2026\/04\/google-gemm4\/en\/resources\/1gemma-4__elo-score__eval__dark_Web-1776305647362.png\" style=\"width: 800px; height: 450px;\" rel=\"share\"\/><\/p>\n<p>Source: <a href=\"https:\/\/blog.google\/innovation-and-ai\/technology\/developers-tools\/gemma-4\/\" rel=\"nofollow noopener\" target=\"_blank\">Google blog<\/a><\/p>\n<p>Reactions in the open-model community have focused less on raw scores and more on usability and new licensing. Sam Witteveen <a href=\"https:\/\/www.linkedin.com\/posts\/samwitteveen_gemma4-opensource-ai-activity-7445512423782068224-WH7R?utm_source=share&amp;utm_medium=member_desktop&amp;rcm=ACoAAAABpJcBWvAKfIas8vYBdUCFJnBNf1rtJIo\" rel=\"nofollow noopener\" target=\"_blank\">applauded the Apache 2.0 license<\/a>.<\/p>\n<p>&#13;<\/p>\n<p>This is an actual real Apache 2 license, which means for the first time, you can take Google&#8217;s best open model, modify it, fine tune it, deploy it commercially, do whatever you want with it. No strings attached<\/p>\n<p>&#13;<\/p>\n<p>Nathan Lambert argues that <a href=\"https:\/\/www.interconnects.ai\/p\/gemma-4-and-what-makes-an-open-model\" rel=\"nofollow noopener\" target=\"_blank\">Gemma 4\u2019s value lies in its frictionless integration<\/a>, noting:<\/p>\n<p>&#13;<\/p>\n<p>Gemma 4\u2019s success is going to be entirely determined by ease of use, to a point where a 5-10% swing on benchmarks wouldn\u2019t matter at all. It\u2019s strong enough, small enough, with the right license, and from the U.S., so many companies are going to slot it in.<\/p>\n<p>&#13;<\/p>\n<p>Day-zero distribution is notably broad: weights are available on <a href=\"https:\/\/huggingface.co\/blog\/gemma4\" rel=\"nofollow noopener\" target=\"_blank\">Hugging Face<\/a>, <a href=\"https:\/\/www.kaggle.com\/models\/google\/gemma-4\" rel=\"nofollow noopener\" target=\"_blank\">Kaggle<\/a>, with reference paths through vLLM, <a href=\"https:\/\/github.com\/ggml-org\/llama.cpp\/pull\/21534\" rel=\"nofollow noopener\" target=\"_blank\">llama.cpp<\/a>, <a href=\"https:\/\/ollama.com\/library\/gemma4\" rel=\"nofollow noopener\" target=\"_blank\">Ollama<\/a>, <a href=\"https:\/\/huggingface.co\/collections\/mlx-community\/gemma-4\" rel=\"nofollow noopener\" target=\"_blank\">MLX<\/a>, <a href=\"https:\/\/lmstudio.ai\/models\/gemma-4\" rel=\"nofollow noopener\" target=\"_blank\">LM Studio<\/a>, <a href=\"https:\/\/unsloth.ai\/docs\/models\/gemma-4\" rel=\"nofollow noopener\" target=\"_blank\">Unsloth<\/a>, <a href=\"https:\/\/cookbook.sglang.io\/autoregressive\/Google\/Gemma4\" rel=\"nofollow noopener\" target=\"_blank\">SGLang<\/a>, and <a href=\"https:\/\/build.nvidia.com\/google\/gemma-4-31b-it\" rel=\"nofollow noopener\" target=\"_blank\">NVIDIA NIM<\/a>, plus an <a href=\"https:\/\/huggingface.co\/nvidia\/Gemma-4-31B-IT-NVFP4\" rel=\"nofollow noopener\" target=\"_blank\">NVFP4 quantized 31B checkpoint<\/a> using NVIDIA Model Optimizer. <a href=\"https:\/\/www.kaggle.com\/\" rel=\"nofollow noopener\" target=\"_blank\">Kaggle<\/a> is running <a href=\"https:\/\/www.kaggle.com\/competitions\/gemma-4-good-hackathon\" rel=\"nofollow noopener\" target=\"_blank\">Gemma 4 Good Challenge<\/a>, inviting developers to build products that create meaningful positive change using the new models.<\/p>\n","protected":false},"excerpt":{"rendered":"Google has recently released Gemma 4, a family of open-weight models spanning effective 2B and 4B edge variants,&hellip;\n","protected":false},"author":2,"featured_media":382925,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[6],"tags":[365,15672,170,2797,366,198911,6833,18486,111,139,69,145],"class_list":{"0":"post-382924","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-technology","8":"tag-ai","9":"tag-architecture-design","10":"tag-development","11":"tag-edge-computing","12":"tag-generative-ai","13":"tag-google-gemm4","14":"tag-large-language-models","15":"tag-ml-data-engineering","16":"tag-new-zealand","17":"tag-newzealand","18":"tag-nz","19":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/nz\/wp-json\/wp\/v2\/posts\/382924","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/nz\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/nz\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/nz\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/nz\/wp-json\/wp\/v2\/comments?post=382924"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/nz\/wp-json\/wp\/v2\/posts\/382924\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/nz\/wp-json\/wp\/v2\/media\/382925"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/nz\/wp-json\/wp\/v2\/media?parent=382924"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/nz\/wp-json\/wp\/v2\/categories?post=382924"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/nz\/wp-json\/wp\/v2\/tags?post=382924"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}