Run Any LLM on Your Phone or Any Device - Full Easy Setup Guide



AI Summary

In this rapid-fire hands-on tutorial by Fahad Biza, viewers learn how to run large language models (LLMs) on edge devices such as mobile phones and tablets. The demo is done on Ubuntu 22.04 with an Nvidia RTX 6000 GPU. The tutorial covers system updates, installing the Java Development Kit (JDK) needed for Android emulators, and setting up Android Studio with a virtual device (Pixel 9 Pro). Using Node.js and npm, the tutorial walks through cloning an LLM repo, installing dependencies, and building the application to run on the emulator. The app allows users to select different LLM models and quantization versions for efficient edge inference. The presenter demonstrates the app running locally in the emulator, chatting with a quantized LLaMA model. Source code customization is briefly shown. The video emphasizes ease and speed, enabling deployment of private, local LLMs on edge devices in minutes. Fahad also mentions affordable GPU rentals by Mast Compute and the open-source community Camel AI sponsoring the video.