
LLM Inference guide for Android - Google AI Edge
Apr 1, 2025 · The LLM Inference API lets you run large language models (LLMs) completely on-device for Android applications, which you can use to perform a wide range of tasks, such as …
Running Lightweight LLMs on Android Devices - Medium
Apr 28, 2024 · In this article, we’ll explore how to run small, lightweight models such as Gemma-2B, Phi-2, and StableLM-3B on Android devices 📱 We’ll be utilizing the Tensorflow Lite and …
UbiquitousLearning/mllm: Fast Multimodal LLM on Mobile Devices - GitHub
End-to-end Android app demo; Advanced support: MoE, Prompt Cache, etc.. mllm is a lightweight, fast, and easy-to-use (multimodal) on-device LLM inference engine for mobile …
6 Mobile Apps For Running Local LLMs (Offline) - Sci Fi Logic
Apr 11, 2024 · Check out these 5 apps that could turn your phone into a local LLM powerhouse! Layla! This app lets you run powerful AI right on your phone, without needing internet. It’s …
How to Install and Run LLMs Locally on Android Phones
Apr 22, 2024 · With a simple app, you can now download and run LLM models locally on your Android phone. You can run Phi-2, Gemma, Mistral and Llama models.
LLM on Android with Keras and TensorFlow Lite
Train and deploy your own large language model (LLM) on Android using Keras and TensorFlow Lite. Large language models (LLMs) have revolutionized tasks like text generation and …
Local LLM for Mobile: Run Llama 2 and Llama 3 on Android
Jun 19, 2024 · Learn how to run Llama 2 and Llama 3 on Android with the picoLLM Inference Engine Android SDK. Runs locally on an Android device.
Android OpenAI API compatible mobile app : r/LocalLLaMA - Reddit
Apr 17, 2024 · I'm looking for an app I can use for inference on any LLM stuff that has meets openAI API specs. I tried a few on android with no luck. I'm using llama.cpp. If anyone knows …
GitHub - droidrun/droidrun
DroidRun is a powerful framework for controlling Android devices through LLM agents. It allows you to automate Android device interactions using natural language commands.
DakeQQ/Native-LLM-for-Android - GitHub
Demonstration of running a native Large Language Model (LLM) on Android devices. Currently supported models include: Qwen3: 0.6B, 1.7B, 4B... Qwen2.5-Instruct: 0.5B, 1.5B, 3B...
- Some results have been removed