Safer and Multimodal: Responsible AI with Gemma

3월 12, 2025
Dana Kurniawan Product Manager
Wenjun Zeng Software Engineer
Ryan Mullins Software Engineer

Last year, we released ShieldGemma, a suite of safety content classifier models built on Gemma 2 and designed to detect harmful content in AI models’ text inputs and outputs. As we debut Gemma 3 today, we’re excited to build on our foundation of responsible AI by announcing ShieldGemma 2.

Link to Youtube Video (visible only when JS is disabled)

ShieldGemma 2, built on Gemma 3, is a 4 billion (4B) parameter model that checks the safety of your synthetic and natural images against key categories to help you build robust datasets and models. With this addition to the Gemma family of models, researchers and developers can now easily minimize the risk of harmful content in their models across key areas of harm:

  • Sexually explicit content

  • Dangerous content

  • Violence
Use ShieldGemma as an input filter to any vision language model, or an an output filter of image generation models

We recommend using ShieldGemma 2 as an input filter to vision language models, or as an output filter of image generation systems. ShieldGemma can be used on both synthetic and natural images.


What’s different in ShieldGemma 2?

Moving beyond text, training and understanding image safety in multimodal models brings new challenges, which is why ShieldGemma 2 is built to respond to a wide range of diverse and nuanced styles of imagery.

To train a robust image safety model, we curated training datasets of natural and synthetic images, and instruction-tuned Gemma 3 to demonstrate strong performance. We compared safety policies to the following benchmarks, and will be releasing a technical report that also incorporates third party benchmarks.

ShieldGemma 2 performance
Evaluation results based on optimal F1 score (%, higher is better) on our internal benchmark

Here’s how ShieldGemma can help you build safer AI image applications:

  • Flexibility: Upload any synthetic or natural images, and edit our prompt template to adapt to your needs. Fine-tune on Google Colab or your own GPU.

  • Versatility: All tools that support Gemma 3 support ShieldGemma 2, including popular frameworks like Transformers, JAX, Keras, Ollama, and others.

  • Collaborative: ShieldGemma is open by nature and welcomes community collaborators to keep building inclusively as we collectively push industry safety standards onwards and upwards.

Deploying open models responsibly relies on a whole community effort, and we look forward to exploring how ShieldGemma 2 can be delivered in smaller sizes, across more harm areas, and aligned with multimodal ML Commons taxonomy in the near future.

We’re excited to continue building for safe and responsible multimodal AI!


Get started today

  • Explore ShieldGemma 2 on our developer site, and see our model card for more information.

  • Try ShieldGemma 2 on Google AI Studio, Hugging Face, Ollama, and other platforms.


Team Acknowledgement

Wenjun Zeng, Ryan Mullins, Dana Kurniawan, Yuchi Liu, Mani Malek, Yiwen Song, Dirichi Ike-Njoku, Hamid Palangi, Jindong Gu, Shravan Dheep, Karthik Narashimhan, Tamoghna Saha, Joon Baek, Rick Pereira, Cai Xu, Jingjing Zhou, Aparna Joshi, Will Hawkins