How to Install and Run LLMs Locally on Android Phones

Learn how to bring the power of AI right to your Android phone—no cloud, no internet, just pure on-device intelligence!



How to Install and Run LLMs Locally on Android Phones
Image by Author | Canva

 

Running large language models (LLMs) locally on Android phones means you can access AI models without relying on cloud servers or an internet connection. This local setup ensures privacy by keeping your data secure and on-device. With advancements in mobile hardware, running AI models locally has become a reality. The MLC Chat app makes it easy to experience this powerful technology right on your phone.

This article will explain the significance of running LLMs locally on Android phones and provide a step-by-step tutorial for installing and running them using the MLC Chat app.

 

Why Run LLMs on Android Phones?

 

LLMs are commonly run on cloud servers due to the significant computational power they require. While Android phones have certain limitations in running LLMs, they also open up exciting possibilities.

  • Enhanced Privacy: Since the entire computation happens on your phone, your data stays local, which is crucial for any sensitive information you share.
  • Offline Access: A constant internet connection is not required to access or interact with these models. This is especially useful for users in remote areas or those with limited internet access.
  • Cost Efficiency: Running LLMs on cloud servers involves operational costs like processing power and cloud storage. This approach provides an economical solution for users.

 

Step-by-Step Guide to Install, and Run MLC Chat on Android

 

The MLC Chat App is an application designed to enable users to run and interact with large language models (LLMs) locally on various devices, including mobile phones, without relying on cloud-based services. Follow the steps below to run LLMs locally on an Android device.

 

Step 1: Install the MLC Chat App

First, you need to download the APK for the MLC Chat App(112MB) from the link given below.
MLC Chat App APK File

 

Install the MLC Chat App

 

Once the APK is downloaded, tap on the file to begin installation.

 

Step 2: Download the LLM

After successfully installing the app, open it, and you'll see a list of available LLMs for download. Models of different sizes and capabilities, such as LLama-3.2, Phi-3.5, and Mistral, are available. Select the model according to your needs and tap the download icon next to it to begin the download. For example, since I’m using a mid-range phone like the Redmi Note 10, I opted for a lightweight model like Qwen-2.5 for smoother performance.

 

Download the LLM

 

Step 3: Run the Installed LLM

Once the model is downloaded, a chat icon will appear next to it. Tap the icon to initiate the model.

 

Run the Installed LLM

 

When the model is ready to go, you can start typing prompts and interact with the LLM locally.

 

Run the Installed LLM

 

For example, on a device like the Redmi Note 10, running a smaller model like Qwen2.5 offers a reasonably smooth experience, generating about 1.4 tokens per second. While this performance is slower compared to high-end devices such as the Galaxy S23 Ultra, it remains functional for basic tasks like short conversations and simple content generation.

 

Conclusion

 

Running LLMs locally on Android devices via the MLC Chat app offers an accessible and privacy-preserving way to interact with AI models. The performance depends heavily on your phone's hardware. This solution is ideal for users who need offline access to AI models, experiment with LLMs in real-time, or are concerned about privacy. As mobile hardware continues to improve, the capabilities of local LLMs will only expand, making this an exciting frontier for AI technology.

 
 

Kanwal Mehreen is a machine learning engineer and a technical writer with a profound passion for data science and the intersection of AI with medicine. She co-authored the ebook "Maximizing Productivity with ChatGPT". As a Google Generation Scholar 2022 for APAC, she champions diversity and academic excellence. She's also recognized as a Teradata Diversity in Tech Scholar, Mitacs Globalink Research Scholar, and Harvard WeCode Scholar. Kanwal is an ardent advocate for change, having founded FEMCodes to empower women in STEM fields.


Get the FREE ebook 'KDnuggets Artificial Intelligence Pocket Dictionary' along with the leading newsletter on Data Science, Machine Learning, AI & Analytics straight to your inbox.

By subscribing you accept KDnuggets Privacy Policy


Get the FREE ebook 'KDnuggets Artificial Intelligence Pocket Dictionary' along with the leading newsletter on Data Science, Machine Learning, AI & Analytics straight to your inbox.

By subscribing you accept KDnuggets Privacy Policy

Get the FREE ebook 'KDnuggets Artificial Intelligence Pocket Dictionary' along with the leading newsletter on Data Science, Machine Learning, AI & Analytics straight to your inbox.

By subscribing you accept KDnuggets Privacy Policy

No, thanks!