Welcome to EdgeAI
Welcome to EdgeAI Documentation - the first Android app to demonstrate real Llama3.2-1B inference using ExecuTorch with Qualcomm QNN backend.
What is EdgeAI?
EdgeAI is a groundbreaking Android application that brings real AI inference to mobile devices. Unlike simulated or rule-based approaches, EdgeAI runs actual Llama3.2-1B model inference on your Android device using:
- ExecuTorch: PyTorch's mobile-optimized runtime
- Qualcomm QNN: Hardware acceleration via Qualcomm AI Engine Direct
- v79 Context Binaries: Optimized for SoC Model-69
- ARM64-v8a: Native Android architecture support
Key Features
✅ Real Model Inference - Actual Llama3.2-1B running on device
✅ Hardware Acceleration - Qualcomm QNN backend for optimal performance
✅ Professional Documentation - Comprehensive guides and technical details
✅ Open Source - Full source code available on GitHub
✅ Mobile Optimized - Designed specifically for Android devices
Quick Start
Get started with EdgeAI in just a few steps:
- Setup Qualcomm AI HUB - Configure your development environment
- Understand Project Structure - Learn how EdgeAI is organized
- Check Latest Release - See what's new in the latest version
Architecture Overview
EdgeAI uses a sophisticated architecture combining:
- Android App Layer: Kotlin/Java UI and application logic
- Native C++ Layer: High-performance inference engine
- ExecuTorch Runtime: PyTorch mobile optimization
- Qualcomm QNN Backend: Hardware acceleration
- Model Layer: Llama3.2-1B with optimized context binaries
Getting Help
- 📖 Documentation: Comprehensive guides and technical details
- 🐛 Issues: Report bugs and request features on GitHub
- 💬 Discussions: Join the community discussions
- 📧 Contact: Reach out for support and collaboration
What's Next?
Ready to dive deeper? Check out our technical documentation:
- Technical Documentation - Deep dive into the architecture
- Setup Guides - Step-by-step configuration
- Release Notes - Latest updates and features
Built with ❤️ for the AI community