Google has released an experimental API that allows large language models to run fully on-device across Android, iOS, and web platforms.
Introduced March 7, the MediaPipe LLM Inference API was designed to streamline on-device LLM integration for web developers, and supports web, Android, and iOS platforms. The API provides initial support for four LLMs: Gemma, Phi 2, Falcon, and Stable LM.
Google warns that the API is experimental and still under active development, but gives researchers and developers the ability to prototype and test openly available models on-device. For Android, Google noted that production applications with LLMs can use the Gemini API or Gemini Nano on-device through Android AICore, a system-level capability introduced in Android 14 that provides Gemini-powered solutions for high-end devices including integrations with accelerators, safety filters, and LoRA adapters.
Email us: contact@neptunesolution.in
Call: 0172-4102740, +91-9780373638, 7495055288 for more details.
Visit us: www.neptunesolution.in
Office address: Sector 34-A, SCO 156-157, second floor, Near Verka Corporate Office, Chandigarh – 160022