Introducing Google AI Edge Portal: Benchmark Edge AI at scale. Sign-up to request access during private preview.

High-performance ML & GenAI deployment on edge platforms

Efficient conversion, runtime, and optimization for on-device machine learning.

Built on the battle-tested foundation of TensorFlow Lite

LiteRT isn't just new; it's the next generation of the world's most widely deployed machine learning runtime. It powers the apps you use every day, delivering low latency and high privacy on billions of devices.

Trusted by the most critical Google apps

100K+ applications, billions of global users

LiteRT highlights

Cross platform ready

Unleash genAI

Simplified hardware acceleration

Multi-framework support

Deploy via LiteRT

Obtain a model

Use .tflite pre-trained models or convert PyTorch, JAX or TensorFlow models to .tflite

Optimize the model

Optionally Quantize the model

Run

Pick desired accelerator and run on LiteRT

Samples, models, and demo

See sample app

Tutorials show you how to use LiteRT with complete, end-to-end examples.

See genAI models

Pre-trained, out-of-the-box Gen AI models.
A gallery that showcases on-device ML/GenAI use cases

Except as otherwise noted, the content of this page is licensed under the Creative Commons Attribution 4.0 License, and code samples are licensed under the Apache 2.0 License. For details, see the Google Developers Site Policies. Java is a registered trademark of Oracle and/or its affiliates.

Last updated 2025年12月19日 UTC.