Helping developers build safer AI experiences for teens

OpenAI Blog Products

Summary

OpenAI releases prompt-based safety policies and the open-weight gpt-oss-safeguard model to help developers build age-appropriate AI experiences for teens, covering risks like graphic content, harmful behaviors, and dangerous activities.

OpenAI releases prompt-based teen safety policies for developers using gpt-oss-safeguard, helping moderate age-specific risks in AI systems.
Original Article Export to Word Export to PDF
View Cached Full Text

Cached at: 04/20/26, 02:51 PM

# Helping developers build safer AI experiences for teens Source: [https://openai.com/index/teen-safety-policies-gpt-oss-safeguard/](https://openai.com/index/teen-safety-policies-gpt-oss-safeguard/) Today, we’re releasing prompt\-based[safety policies⁠\(opens in a new window\)](https://github.com/openai/teen-safety-policy-pack)to help developers create age\-appropriate protections for teens\. Built to work with our open\-weight safety model,[gpt\-oss\-safeguard⁠\(opens in a new window\)](https://huggingface.co/openai/gpt-oss-safeguard-20b), these policies simplify how developers turn safety requirements into usable classifiers for real\-world systems\. We released open weight models to democratize access to powerful AI and support broad innovation\. At the same time, we believe safety and innovation go hand in hand, and that developers should have access to capable models as well as the tools and policies to deploy them safely and responsibly\. We developed these policies to support developers in their safety efforts to protect young users, with input from trusted external organizations including[Common Sense Media⁠\(opens in a new window\)](https://www.commonsensemedia.org/)and[everyone\.ai⁠\(opens in a new window\)](http://everyone.ai/)\. We recognize that teens and adults have different needs, and that teens need additional protections\. These policies are designed to help developers account for those differences and build experiences that are both empowering and appropriate for younger users\. Today’s release builds on that foundation\. We’re making these safety policies available to developers to support them in deploying safety protections for teens and helping democratize access across the open weights ecosystem\. While safety classifiers like gpt\-oss\-safeguard can detect harmful content, they depend on clear definitions of what that content is\. In practice, one of the biggest challenges developers face is defining policies that accurately capture teen\-specific risks and can be consistently applied in real systems\. Even experienced teams often struggle to translate high\-level safety goals into precise, operational rules, especially since it requires both subject matter expertise and deep AI knowledge\. This can lead to gaps in protection, inconsistent enforcement, or overly broad filtering\. Clear, well\-scoped policies are a critical foundation for effective safety systems\. To address this challenge, we are releasing a set of[safety policies⁠\(opens in a new window\)](https://github.com/openai/teen-safety-policy-pack), tailored to common risks faced by teens and informed by careful review of existing research about teens’ unique developmental differences\. These policies are structured as prompts that can be directly used with[gpt\-oss\-safeguard⁠\(opens in a new window\)](https://huggingface.co/openai/gpt-oss-safeguard-20b)and other reasoning models, enabling developers to more easily apply consistent safety standards across their systems\. The initial release includes policies covering: - Graphic violent content - Graphic sexual content - Harmful body ideals and behaviors - Dangerous activities and challenges - Romantic or violent roleplay - Age\-restricted goods and services These policies can be used for real\-time content filtering, as well as offline analysis of user\-generated content\. By structuring policies as prompts, developers can more easily integrate them into existing workflows, adapt them to their use cases, and iterate over time\. We worked with external organizations including[Common Sense Media⁠\(opens in a new window\)](https://www.commonsensemedia.org/)and[everyone\.ai⁠\(opens in a new window\)](http://everyone.ai/)to inform the development of these policies\. Their expertise helped shape the scope of content to cover, strengthen the structure of the prompts, and refine the edge cases to consider when evaluating them\. This work reflects an ongoing effort to collaborate with experts and the broader ecosystem to improve how AI systems support young people\. *“One of the biggest gaps in AI safety for teens has been the lack of clear, operational policies that developers can build from\. Many times, developers are starting from scratch\. These prompt\-based policies help set a meaningful safety floor across the ecosystem, and because they're released as open source, they can be adapted and improved over time\. We're encouraged to see this kind of infrastructure being made available broadly, and we hope it catalyzes more shared youth\-safety starting points across the industry\.”* —**Robbie Torney, Head of AI & Digital Assessments, Common Sense Media** *“Efforts like this that make youth safety policies more operational are valuable because they help translate expert knowledge into guidance that can be used in real systems\. Content policies are an important first step, and they also open the door to broader work on how model behavior can shape youth\-relevant risks over time\. Inspired by this work and our own research,*[*everyone\.ai*⁠\(opens in a new window\)](http://everyone.ai/)*has also created an initial behavioral policy focused on risks like exclusivity and overreliance\."* *—***Dr\. Mathilde Cerioli, Chief Scientist at everyone\.AI** The policies are intended as a starting point, not as a comprehensive or final definition or guarantee of teen safety\. Each application has unique risks, audiences and contexts, and developers are best positioned to understand the risks that their products and AI integrations may present\. We strongly encourage developers to adapt and extend these policies based on their specific needs and combine them with other safeguards such as product design decisions, user controls, teen\-friendly transparency, monitoring systems and thoughtful, age\-appropriate responses\. We believe a layered[defense in depth⁠⁠](https://openai.com/safety/how-we-think-about-safety-alignment/#defense-in-depth)approach is essential to building safer AI systems\. These policies draw from our internal experience, but they do not reflect the full extent of OpenAI’s internal policies or safeguards\. Developers and organizations can adapt these policies to their specific applications, translate them into different languages, and extend them to cover additional risk areas\. Over time, we hope this contributes to a more robust and shared foundation for implementing safety policies in AI systems\.

Similar Articles

Introducing the Teen Safety Blueprint

OpenAI Blog

OpenAI introduces the Teen Safety Blueprint, a comprehensive framework for building AI tools that protect and empower teenagers, including age-appropriate design, product safeguards, and parental controls. The initiative demonstrates OpenAI's proactive approach to teen safety with implementations like parental controls, age-prediction systems, and strengthened safeguards across products.

Teen safety, freedom, and privacy

OpenAI Blog

OpenAI outlines its approach to balancing teen safety, user freedom, and privacy in ChatGPT, including building an age-prediction system, parental controls, and stricter content rules for under-18 users. The company also signals plans for advanced privacy features and advocates for AI conversation privilege with policymakers.

Updating our Model Spec with teen protections

OpenAI Blog

OpenAI has updated its Model Spec with new Under-18 Principles to guide ChatGPT's behavior for teen users aged 13-17, focusing on safety, age-appropriate interactions, and stronger guardrails around high-risk topics like self-harm and explicit content. The update was developed with input from the American Psychological Association and is grounded in developmental science.

OpenAI’s commitment to child safety: adopting safety by design principles

OpenAI Blog

OpenAI and major tech companies including Amazon, Google, Meta, and Microsoft have committed to implementing Safety by Design principles for child protection in generative AI development, deployment, and maintenance. The initiative aims to mitigate risks of child sexual abuse material generation and spread through comprehensive measures across model development, release, and ongoing platform safety.