Google logo

Gemma 3n

GoogleOpen SourceVerified

Gemma 3n is a mobile-first AI model built on a groundbreaking architecture optimized for on-device performance. It features Per-Layer Embeddings (PLE) for reduced RAM usage, allowing larger models to run on mobile devices with a memory footprint comparable to much smaller models. Supports multimodal understanding including text, images, audio, and video.

2025-05-20
4B
Mobile-optimized Transformer with Per-Layer Embeddings (PLE)
Proprietary

Specifications

Parameters
4B
Architecture
Mobile-optimized Transformer with Per-Layer Embeddings (PLE)
License
Proprietary
Context Window
32,000 tokens
Type
multimodal
Modalities
textimageaudiovideo

Benchmark Scores

MMLU62.2

Massive Multitask Language Understanding (MMLU) tests knowledge across 57 subjects including mathema...

WMT2450.1

WMT24 is the 2024 edition of the Workshop on Machine Translation, which provides a ranking of Genera...

Advanced Specifications

Model Family
Gemma
API Access
Available
Chat Interface
Available
Multilingual Support
Yes
Hardware Support
Mobile SoCsQualcommMediaTekSamsung System LSI

Capabilities & Limitations

Capabilities
on-device inferencemultimodal understandingaudio transcriptiontranslationmix'n'match submodelsoffline operation
Known Limitations
Limited to mobile-optimized tasksReduced capabilities compared to cloud models
Notable Use Cases
on-device AI assistantsprivate speech recognitionoffline translationreal-time visual understanding
Tool Use Support
No

Related Models