[Hands-on] Deploy and Run LLMs on your Phone!
Fine-tune and deploy LLMs directly on iOS and Android devices using UnslothAI, TorchAO, and ExecuTorch. The tutorial walks through loading Qwen3-0.6B, preparing reasoning and chat datasets, training with quantization-aware methods, exporting to mobile-ready .pte format, and running the model locally on iPhone at ~25 tokens/second. The resulting model is ~470MB and runs 100% on-device without requiring cloud connectivity.