🔥 google-ai-edge/LiteRT-LM
google-ai-edge/LiteRT-LM is trending on GitHub today with 113 new stars.
LiteRT-LM is Google's production-ready, high-performance, open-source inference framework for deploying Large Language Models on edge devices.
🔗 Product Website
🔥 What's New: Gemma 4 support with LiteRT-LM
Deploy Gemma 4 across a broad range of hardware with stellar performance (blog).
👉 Try on Linux, macOS, Windows (WSL) or Raspberry Pi with the LiteRT-LM CLI:
litert-lm run \ --from-huggingface-repo=litert-community/gemma-4-E2B-it-litert-lm \ gemma-4-E2B-it.litertlm \ --prompt="What is the capital of France?"litert-lm run \ --from-huggingface-repo=litert-community/gemma-4-E2B-it-litert-lm \ gemma-4-E2B-it.litertlm \ --prompt="What is the capital of France?"🌟 Key Features
-
📱 Cross-Platform Support: Android, iOS, Web, Desktop, and IoT (e.g. Raspberry Pi).
-
🚀 Hardware Acceleration: Peak performance via GPU and NPU accelerators.
-
👁️ Multi-Modality: Support for vision and audio inputs.
-
🔧 Tool Use: Function calling support for agentic workflows.
-
📚 Broad Model Support: Gemma, Llama, Phi-4, Qwen, and more.
🚀 Production-Ready for Google's Products
LiteRT-LM powers on-device GenAI experiences in Chrome, Chromebook Plus, Pixel Watch, and more.
You can also try the Google AI Edge Gallery app to run models immediately on your device.
Install the app today from Google Play Install the app today from App Store
📰 Blogs & Announcements
Link Description
Bring state-of-the-art agentic skills to the edge with Gemma 4 Deploy Gemma 4 in-app and across a broader range of devices with stellar performance and broad reach using LiteRT-LM.
On-device GenAI in Chrome, Chromebook Plus and Pixel Watch Deploy language models on wearables and browser-based platforms using LiteRT-LM at scale.
On-device Function Calling in Google AI Edge Gallery Explore how to fine-tune FunctionGemma and enable function calling capabilities powered by LiteRT-LM Tool Use APIs.
Google AI Edge small language models, multimodality, and function calling Latest insights on RAG, multimodality, and function calling for edge language models.
🏃 Quick Start
🔗 Key Links
-
👉 Technical Overview including performance benchmarks, model support, and more.
-
👉 LiteRT-LM CLI Guide including installation, getting started, and advanced usage.
⚡ Quick Try (No Code)
Try LiteRT-LM immediately from your terminal without writing a single line of code using uv:
uv tool install litert-lm
litert-lm run
--from-huggingface-repo=google/gemma-3n-E2B-it-litert-lm
gemma-3n-E2B-it-int4
--prompt="What is the capital of France?"`
📚 Supported Language APIs
Ready to get started? Explore our language-specific guides and setup instructions.
Language Status Best For... Documentation
Kotlin ✅ Stable Android apps & JVM Android (Kotlin) Guide
Python ✅ Stable Prototyping & Scripting Python Guide
C++ ✅ Stable High-performance native C++ Guide
Swift 🚀 In Dev Native iOS & macOS (Coming Soon)
🏗️ Build From Source
This guide shows how you can compile LiteRT-LM from source.
📦 Releases
-
v0.9.0: Improvements to function calling capabilities, better app performance stability.
-
v0.8.0: Desktop GPU support and Multi-Modality.
-
v0.7.0: NPU acceleration for Gemma models.
For a full list of releases, see GitHub Releases.
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Open Source AI

Qwen 3.5 Tool Calling Fixes for Agentic Use: What's Broken, What's Fixed, What You (may) Still Need
Posted - What follows after this introduction is generated by Claude Opus 4.6 after hundreds of back and forths with log analysis for tool calls that were not working, and Qwen 3.5 models getting confused from local llm providers as well as Nano-Gpt. I fixed it for my own use with Pi coding agent at the time. Some of the fixes that were needed are no longer needed (TLDR at the bottom) but most are still applicable, as validated today. If you use Qwen 3.5 models and are having issues with model performance, tool calls, or general instability, the reference below might be a useful read. In the end, the fixes below on pi coding agent + llamacpp + Bartowski's quants (for stability) is what took my experience to 99% reliability and quality with all Qwen 3.5 models (Q5_k_L). Hope it helps someon

NLP SOTA By far, (NOT WORK ON CNN) , Can save 66% FLOPs and same or BETTER accuracy than baseline
NLP SOTA. CNN NOT WORK, IN NLP WE REDUCE 66% FLOPs with same or MORE accuracy, you have the link of github on zenodo page. DeepFocus-BP: Error-Aware Adaptive Backpropagation via Dynamic Alpha-Beta Routing (Achieving 66% FLOPs Reduction with Improved Accuracy) - SOTA NLP Confirmed v3. (Resnet FAIL) 3 posts - 2 participants Read full topic


Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!