LiteRT
Google’s high-performance runtime for on-device AI (formerly TensorFlow Lite)
Cross‑platform on‑device ML
Leverage LiteRT and Qualcomm AI Hub to run powerful machine learning models across devices, optimized for on-device machine learning with multi-platform support.
Submit on AI Hub
To use LiteRT when submitting your compile job to Qualcomm AI Hub, please specify ‑‑target_runtime tflite
.
Bringing ML‑powered experiences to over 100K apps running on 2.7B devices
Visit LiteRT