Anahtar Kelimeler:Gemma 3n, Uçta çok modlu, MatFormer, Katman katman gömme, Düşük kaynak tüketimi, Gemma 3n E2B modeli, Gemma 3n E4B modeli, LMArena puanı, 2GB RAM’de çalıştırma, Hugging Face’de yayınlandı
🔥 Focus
Google Releases Gemma 3n, Ushering in a New Era of On-Device Multimodality: Google has officially released the Gemma 3n model series, designed for on-device use with native support for text, image, audio, and video inputs. The series includes two models, E2B and E4B. Although their actual parameter counts are 5B and 8B, thanks to the innovative MatFormer “Matryoshka doll” architecture and Per-Layer Embedding (PLE) technology, their memory footprint is equivalent to that of 2B and 4B models, capable of running on as little as 2GB of RAM. Gemma 3n scored over 1300 on the LMArena leaderboard, making it the first model under 10B parameters to achieve this score, showcasing its excellent performance with low resource consumption. The model is now fully available on major open-source platforms like Hugging Face, Ollama, and MLX, driving further development of on-device AI applications. (Source: HuggingFace Blog, karminski3, demishassabis, Reddit r/LocalLLaMA)