Gemma 3n : Google reduces the size of cutting-edge artificial intelligence

Publié le 24 May 2025 à 09h42
modifié le 24 May 2025 à 09h42

Gemma 3n redefines the standards of artificial intelligence with a significant reduction in size. Google revolutionizes inference with its extremely efficient model designed for devices. An innovative architecture breaks performance barriers in the face of limited resources. This multimodal SLM combines text, audio, video, and images while maintaining remarkable efficiency. The transformations brought about by Gemma 3n could redefine our interaction with technology, simplifying access to advanced artificial intelligence.

Introduction to Gemma 3n

Google recently unveiled the Gemma 3n model, an innovative multimodal language system (SLM), at Google I/O 2025. This model, developed by the DeepMind team, is distinguished by its ability to process data in various forms such as text, audio, video, and images. Its design has been optimized for inference on CPU, making it accessible on devices with limited resources.

An Innovative Architecture

The Gemma model family incorporates technological advancements from its predecessor, Gemini. DeepMind engineers have taken a radical approach by developing a new architecture dedicated to use on less powerful devices. The major innovation, named Per-Layer, significantly reduces RAM consumption. Thus, Gemma 3n, equipped with 5 or 8 billion parameters, operates with a memory footprint much lower than that of similar models.

Performance and Benchmarks

On reference platforms like the Chatbot Arena, Gemma 3n achieves an impressive Elo score of 1269, placing it just behind Claude 3.7 Sonnet. The performance is all the more remarkable for a model of this size. Results on traditional benchmarks, such as 64.9% on MMLU and 63.6% on MBPP, confirm its status as a model of excellence.

Technical Specificities

The MatFormer, another innovation of the architecture, allows for the integration of a sub-model of 2 billion parameters. This functionality helps to adjust model sizes according to task complexity. Developers can thus recreate various sizes of sub-models, maximizing the efficiency of the resources used.

Accessibility and Use

Gemma 3n is already available via Google AI Studio at no cost, and users can also download the model weights on Hugging Face. Currently, the deployed version only allows for processing text and image modalities, but updates are underway to integrate all modalities.

Terms of Use

Using this model for commercial purposes incurs no licensing fees or royalties for Google. However, certain restrictions apply. The use of Gemma 3n is prohibited for generating protected or illegal content. Automated decision-making in areas affecting individual rights, such as finance or healthcare, is also prohibited.

Recommended Applications

Gemma 3n sets a new standard in the field of open-source SLMs. Google recommends its integration for text generation, information summarization, visual analysis, and audio transcription. A notable feature is its optimization for mobile inference, with a RAM requirement limited to only 3924 MB, making it ideal for exploring new uses, as mentioned in these projects: Reachy 2, OpenAI, and AI in business.

Conclusion on Its Superiority

Gemma 3n juxtaposes performance and modularity in a compact form. This model, in line with the latest advancements in artificial intelligence, embodies a precise response to the growing demand for efficiency in SLMs. Its reduced size contrasts with its impressive results on specific benchmarks, allowing it to position itself at the forefront of the technological competition.

User FAQ on Gemma 3n: Google Reduces the Size of Cutting-Edge Artificial Intelligence

What is Gemma 3n and how does it differ from other artificial intelligence models?
Gemma 3n is a multimodal artificial intelligence model developed by Google, designed to work efficiently on devices with limited hardware capabilities. Its main innovation is the Per-Layer architecture, which optimizes RAM consumption while maintaining very good performance on various benchmarks.

How does Gemma 3n manage to reduce its memory footprint?
The Per-Layer Embeddings technique used in Gemma 3n allows for dynamic reduction of RAM usage by optimizing the representations of each layer, resulting in the model achieving performances similar to those of models with fewer parameters.

What types of data can Gemma 3n process?
Gemma 3n is fully multimodal and designed to process text, audio, video, and images, although the current version primarily focuses on text and image modalities. Future updates should expand its capabilities.

What is Gemma 3n’s performance score compared to other models?
On the Chatbot Arena, Gemma 3n achieves an Elo score of 1269, placing it just behind Claude 3.7 Sonnet and ahead of other models like GPT-4.1. Additionally, it delivers impressive results on classic benchmarks such as MMLU and HumanEval.

Is Gemma 3n available as open source and what are the terms of use?
Yes, Gemma 3n is available as open-source. Users can use it for commercial purposes without licensing fees, but Google reserves the right to restrict its use if it violates its terms of use, particularly for content protected by copyright.

What are the recommended practical applications for Gemma 3n?
Gemma 3n is recommended for various applications such as text generation, chatbot use, information summarization, as well as visual analysis and audio file transcription, thanks to its reduced size and optimization for mobile inference.

How can developers customize Gemma 3n according to their needs?
Developers can create multiple sizes of sub-models in Gemma 3n thanks to the MatFormer architecture, which allows for the native integration of a sub-model optimized according to the complexity of each task, thereby reducing resource requirements.

actu.iaNon classéGemma 3n : Google reduces the size of cutting-edge artificial intelligence

Disney and NBC Universal are suing the artificial intelligence image generator Midjourney

découvrez comment l'union européenne mobilise 145,5 millions d'euros pour améliorer la cybersécurité des petites et moyennes entreprises, des hôpitaux et des administrations publiques, renforçant ainsi la sécurité numérique et la résilience de ces secteurs essentiels.

Money in the service of the mission at Wikimedia: interview with Lane Becker

découvrez l'interview de lane becker sur l'importance de l'argent dans la mission de wikimedia. plongez dans les stratégies et les visions qui guident l'organisation pour rendre l'information accessible à tous.

Create a French-speaking voice agent with Rounded: our simplified experience

découvrez comment créer facilement un agent vocal francophone grâce à rounded. notre expérience simplifiée vous guide pas à pas pour concevoir une solution personnalisée qui répond à vos besoins.

The impact of artificial intelligence on 3D animation training at ESMA

découvrez comment l'intelligence artificielle transforme les formations en animation 3d à l'esma, en enrichissant les compétences des étudiants et en révolutionnant les méthodes d'enseignement. explorez les nouvelles technologies et leur impact sur le processus créatif et l'industrie de l'animation.

the impact of artificial intelligence on arbitration at the Fifa Club World Cup

découvrez comment l'intelligence artificielle révolutionne l'arbitrage lors de la coupe du monde des clubs de la fifa. analyse des technologies innovantes, de leur impact sur la précision des décisions et des implications pour l'avenir du football.