FunctionGemma released, a model tuned for function calling! Learn more

ShieldGemma

ShieldGemma is a set of instruction tuned models for evaluating the safety of text and images against a set of defined safety policies. You can use this model as part of a larger implementation of a generative AI application to help evaluate and prevent generative AI applications from violating safety policies. The ShieldGemma family of models is provided with open weights to allow you to fine-tune it for your specific use case.

ShieldGemma 2 is a 4B parameter model built to label images for safety.

ShieldGemma 1 is built on Gemma 2 in 2B, 9B, and 27B parameter sizes.

  • Content safety evaluation

    Evaluate the safety of prompt input and output responses against a set of defined safety policies.
  • Tuneable, open models

    ShieldGemma models are provided with open weights and can be fine-tuned for your specific use case.

ShieldGemma 2

View the model card

ShieldGemma's 2 model card contains detailed information about the model implementation, evaluations, model usage and limitations, and more.

View on Kaggle

View more code, Colab notebooks, information, and discussions about ShieldGemma on Kaggle.

View on HuggingFace

Run a working example for using ShieldGemma to evaluate images.

ShieldGemma 1

View the model card

ShieldGemma's model card contains detailed information about the model implementation, evaluations, model usage and limitations, and more.

View on Kaggle

View more code, Colab notebooks, information, and discussions about ShieldGemma on Kaggle.

Run in Colab

Run a working example for using ShieldGemma to evaluate text prompt input and output.

Except as otherwise noted, the content of this page is licensed under the Creative Commons Attribution 4.0 License, and code samples are licensed under the Apache 2.0 License. For details, see the Google Developers Site Policies. Java is a registered trademark of Oracle and/or its affiliates.

Last updated 2025年06月02日 UTC.