{"id":813049,"date":"2025-07-01T20:53:58","date_gmt":"2025-07-01T12:53:58","guid":{"rendered":"https:\/\/ztylezman.com\/?p=813049"},"modified":"2025-07-05T10:10:32","modified_gmt":"2025-07-05T02:10:32","slug":"google-launches-gemma-3n-multimodal-ai-on-mobile-with-2gb-memory","status":"publish","type":"post","link":"https:\/\/ztylezman.com\/en\/gadgets-en-2\/google-launches-gemma-3n-multimodal-ai-on-mobile-with-2gb-memory\/","title":{"rendered":"Google Launches Gemma 3n Multimodal AI Model Running on Mobile Devices with 2GB Memory"},"content":{"rendered":"\n<p>Google has recently launched its brand-new multimodal open-source AI model, Gemma 3n. The standout feature of this model is its ability to run locally on smartphones with just 2GB of memory, offering developers a seamless AI application experience. Gemma 3n supports not only text input but also voice, image, and video data processing, enabling users to unlock full multimodal capabilities on mobile devices.<\/p>\n<p>As part of the Google Gemma model family, the development philosophy of Gemma 3n consistently emphasizes openness and customization. The officially released E4B main model and E2B sub-model are both available for download on the Hugging Face and Kaggle platforms. These models are structured using the &#8220;MatFormer&#8221; technology, resembling a Russian doll in their architecture, with E2B serving as a streamlined version of E4B that can operate independently.<\/p>\n<p>The E2B model, designed specifically for mobile platforms, boasts 5 billion parameters and successfully compresses memory usage to just 2GB. This achievement is thanks to its innovative &#8220;Per-Layer Embeddings (PLE)&#8221; design, which significantly reduces the data access burden during model execution, allowing memory-intensive models to run smoothly on smartphones.<\/p>\n<p>It\u2019s worth mentioning that the Gemma 3n E4B model boasts 8 billion parameters, and in practice, its memory consumption during operation is comparable to traditional 4 billion models, requiring only 3GB. In terms of performance, the Gemma 3n E4B even surpasses the GPT 4.1-nano and outperforms larger models like the Llama 4 Maverick 17B-128E and Phi-4.<\/p>\n<p>Currently, Gemma 3n supports multiple execution environments, including Hugging Face Transformers, llama.cpp, Google AI Edge, Ollama, and MLX. Users can also run the Google AI Edge Gallery application locally on mobile devices like the Pixel 8 Pro. Additionally, users can test its chat functionality in Google AI Studio.<\/p>\n<p>Google has showcased the potential of AI models freed from hardware limitations through Gemma 3n. This design, which allows for open downloads and low memory requirements, could drive the rapid adoption of multimodal AI applications in smartphones and IoT devices in the near future.<\/p>\n\n","protected":false},"excerpt":{"rendered":"<p>Google&#8217;s Gemma 3n is a new open-source multimodal AI model that operates locally on smartphones with only 2GB RAM, supporting text voice images and videos, marking a breakthrough in mobile AI capabilities.<\/p>\n","protected":false},"author":9,"featured_media":807326,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"inline_featured_image":false,"footnotes":"Google has recently launched its advanced multimodal open-source AI model, Gemma 3n, designed to run locally on smartphones with just 2GB of memory. This model supports diverse data types including text, voice, images, and videos, and is part of the Google Gemma family emphasizing openness and customization. The E4B and E2B models, with 8 and 5 billion parameters respectively, leverage MatFormer technology, making high-performance AI accessible on mobile platforms. Compatible with Hugging Face Transformers, llama.cpp, Google AI Edge, Ollama, and MLX environments, Gemma 3n demonstrates potential to accelerate multimodal AI adoption in smartphones and IoT devices by enabling efficient, hardware-independent AI processing."},"categories":[5012],"tags":[],"class_list":{"0":"post-813049","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-gadgets-en-2"},"raw_content":"<!-- wp:html \/-->\n<!-- wp:paragraph --><p>Google has recently launched its brand-new multimodal open-source AI model, Gemma 3n. The standout feature of this model is its ability to run locally on smartphones with just 2GB of memory, offering developers a seamless AI application experience. Gemma 3n supports not only text input but also voice, image, and video data processing, enabling users to unlock full multimodal capabilities on mobile devices.<\/p><!-- \/wp:paragraph -->\n<!-- wp:paragraph --><p>As part of the Google Gemma model family, the development philosophy of Gemma 3n consistently emphasizes openness and customization. The officially released E4B main model and E2B sub-model are both available for download on the Hugging Face and Kaggle platforms. These models are structured using the \"MatFormer\" technology, resembling a Russian doll in their architecture, with E2B serving as a streamlined version of E4B that can operate independently.<\/p><!-- \/wp:paragraph -->\n<!-- wp:paragraph --><p>The E2B model, designed specifically for mobile platforms, boasts 5 billion parameters and successfully compresses memory usage to just 2GB. This achievement is thanks to its innovative \"Per-Layer Embeddings (PLE)\" design, which significantly reduces the data access burden during model execution, allowing memory-intensive models to run smoothly on smartphones.<\/p><!-- \/wp:paragraph -->\n<!-- wp:paragraph --><p>It\u2019s worth mentioning that the Gemma 3n E4B model boasts 8 billion parameters, and in practice, its memory consumption during operation is comparable to traditional 4 billion models, requiring only 3GB. In terms of performance, the Gemma 3n E4B even surpasses the GPT 4.1-nano and outperforms larger models like the Llama 4 Maverick 17B-128E and Phi-4.<\/p><!-- \/wp:paragraph -->\n<!-- wp:paragraph --><p>Currently, Gemma 3n supports multiple execution environments, including Hugging Face Transformers, llama.cpp, Google AI Edge, Ollama, and MLX. Users can also run the Google AI Edge Gallery application locally on mobile devices like the Pixel 8 Pro. Additionally, users can test its chat functionality in Google AI Studio.<\/p><!-- \/wp:paragraph -->\n<!-- wp:paragraph --><p>Google has showcased the potential of AI models freed from hardware limitations through Gemma 3n. This design, which allows for open downloads and low memory requirements, could drive the rapid adoption of multimodal AI applications in smartphones and IoT devices in the near future.<\/p><!-- \/wp:paragraph -->\n\n<!-- wp:html \/-->","_links":{"self":[{"href":"https:\/\/ztylezman.com\/en\/wp-json\/wp\/v2\/posts\/813049","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/ztylezman.com\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/ztylezman.com\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/ztylezman.com\/en\/wp-json\/wp\/v2\/users\/9"}],"replies":[{"embeddable":true,"href":"https:\/\/ztylezman.com\/en\/wp-json\/wp\/v2\/comments?post=813049"}],"version-history":[{"count":0,"href":"https:\/\/ztylezman.com\/en\/wp-json\/wp\/v2\/posts\/813049\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/ztylezman.com\/en\/wp-json\/wp\/v2\/media\/807326"}],"wp:attachment":[{"href":"https:\/\/ztylezman.com\/en\/wp-json\/wp\/v2\/media?parent=813049"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/ztylezman.com\/en\/wp-json\/wp\/v2\/categories?post=813049"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/ztylezman.com\/en\/wp-json\/wp\/v2\/tags?post=813049"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}