Gemma 3n
Gemma 3n is a mobile-first AI model built on a groundbreaking architecture optimized for on-device performance. It features Per-Layer Embeddings (PLE) for reduced RAM usage, allowing larger models to run on mobile devices with a memory footprint comparable to much smaller models. Supports multimodal understanding including text, images, audio, and video.
2025-05-20
4B
Mobile-optimized Transformer with Per-Layer Embeddings (PLE)
Proprietary
Specifications
- Parameters
- 4B
- Architecture
- Mobile-optimized Transformer with Per-Layer Embeddings (PLE)
- License
- Proprietary
- Context Window
- 32,000 tokens
- Type
- multimodal
- Modalities
- textimageaudiovideo
Advanced Specifications
- Model Family
- Gemma
- API Access
- Available
- Chat Interface
- Available
- Multilingual Support
- Yes
- Hardware Support
- Mobile SoCsQualcommMediaTekSamsung System LSI
Capabilities & Limitations
- Capabilities
- on-device inferencemultimodal understandingaudio transcriptiontranslationmix'n'match submodelsoffline operation
- Known Limitations
- Limited to mobile-optimized tasksReduced capabilities compared to cloud models
- Notable Use Cases
- on-device AI assistantsprivate speech recognitionoffline translationreal-time visual understanding
- Tool Use Support
- No