Run AI inference on Cloud Run with GPUs

Use GPUs to run AI inference on Cloud Run. If you are new to AI concepts, see GPUs for AI. GPUs are used to train and run AI models. This can give you more stable performance with the ability to scale workloads depending on your overall utilization. See GPU support for services, jobs, and worker pools to learn more about GPU configurations.

Tutorials for services

Tutorials for jobs

Except as otherwise noted, the content of this page is licensed under the Creative Commons Attribution 4.0 License, and code samples are licensed under the Apache 2.0 License. For details, see the Google Developers Site Policies. Java is a registered trademark of Oracle and/or its affiliates.

Last updated 2025年11月20日 UTC.