Google DeepMind's Chintan Parikh presents the Gemma 4 edge models (2B and 4B parameter variants) and the LiteRT on-device inference framework. Key topics include benefits of edge AI deployment (latency, privacy, cost), new Gemma 4 capabilities like function calling, structured JSON output, and chain-of-thought reasoning, plus a demo gallery app showcasing on-device agent skills. The LiteRT framework supports cross-platform deployment across Android, iOS, macOS, Linux, Windows, Raspberry Pi, and IoT devices, with NPU acceleration from Qualcomm and MediaTek delivering up to 13x performance boosts. Models are available on Hugging Face under Apache 2.0 and support PyTorch and JAX conversion paths.
•23m watch time
Sort: