OpenAI said Tuesday it is releasing a set of prompts that developers can use to make their apps safer for teens. The AI lab said the set of teen safety policies can be used with its open-weight safety model known as gpt-oss-safeguard.
Rather than working from scratch to figure out how to make AI safer for teens, developers can use these prompts to fortify what they build. They address issues like graphic violence and sexual content, harmful body ideals and behaviors, dangerous activities and challenges, romantic or violent role play, and age-restricted goods and services.
These safety policies are designed as prompts, making them easily compatible with other models besides gpt-oss-safeguard, though they’re probably most effective within OpenAI’s own ecosystem.
To write these prompts, OpenAI said it worked with AI safety watchdogs, Common Sense Media and everyone.ai.
“These prompt-based policies help set a meaningful safety floor across the ecosystem, and because they’re released as open source, they can be adapted and improved over time,” said Robbie Torney, Head of AI & Digital Assessments at Common Sense Media, in a statement.
OpenAI noted in its blog that developers, including experienced teams, often struggle to translate safety goals into precise, operational rules.
“This can lead to gaps in protection, inconsistent enforcement, or overly broad filtering,” the company wrote. “Clear, well-scoped policies are a critical foundation for effective safety systems.”
Techcrunch event
San Francisco, CA
|
October 13-15, 2026
OpenAI admits that these policies aren’t a solution to the complicated challenges of AI safety. But it builds off its previous efforts, including product-level safeguards such as parental controls and age prediction. Last year, OpenAI updated guidelines for its large language models — known as Model Spec — to tackle how its AI models should behave with users under 18.
OpenAI doesn’t have the cleanest track record itself, however. The company is facing several lawsuits filed by the families of people who died by suicide after extreme ChatGPT use. These dangerous relationships often form after the user eclipses the chatbot’s safeguards, and no model’s guardrails are fully impenetrable. Still, these policies are at least a step forward, especially since it can help indie developers.


