Published on March 24, 2026

OpenAI Sora Safety: Addressing Fakes, Deepfakes, and Child Protection

How OpenAI Is Building Safety into Sora: From Fakes to Child Protection

OpenAI has explained how the safety system in the Sora video generator and its associated application works – from moderation to digital credentials on content.

Security 4 – 6 minutes min read
Event Source: OpenAI 4 – 6 minutes min read

Generating video from a text prompt is one of the most impressive capabilities of modern AI. However, it also raises serious concerns regarding fakes, deepfakes, content depicting children, and disinformation. OpenAI understands these challenges and is proactively discussing how protection is built into Sora and its new application.

Why AI Video Presents Unique Safety Challenges

Why Video Is a Special Case

Text and even image models have already established certain best practices in safety. With video, everything is more complex: it is perceived as more “real,” spreads faster, and is harder to verify. A fake video of a real person can cause much more harm than a text statement attributed to them.

This is precisely why OpenAI emphasizes that Sora was created with safety in mind from the very beginning, rather than as an add-on to a finished product. Simply put, their approach is not “release first, figure it out later,” but “figure it out first.”

Sora Content Generation Restrictions and Prohibited Use Cases

What Is Specifically Prohibited

The list of restrictions is quite predictable, but the details are important. Sora will not generate:

  • sexual content involving minors (this is an absolute restriction, with no exceptions);
  • realistic videos of real people without their consent – especially in compromising or sexual contexts;
  • content that could be used for mass disinformation or manipulation;
  • scenes with real violence presented in a positive light.

At the same time, OpenAI notes that creative freedom is important, and the system should not block everything “just in case.” The balance between what is permissible and what is forbidden is one of the key challenges the team is working on.

Video Provenance: Digital Watermarks to Identify AI-Generated Content

Digital Credentials: How to Distinguish AI Video from Real Video

One of the specific protection mechanisms is digital watermarks. Every video created through Sora receives an invisible marker based on the C2PA standard (an open industry standard for labeling AI-generated content). This marker is embedded in the file itself and is preserved during distribution.

While this is not a panacea – technically, someone could try to remove the marker – in combination with other measures, it allows platforms and verification tools to determine the video's origin. If you see a clip and want to know if it was made by a neural network, the presence of such a marker can provide the answer.

OpenAI Sora Moderation Policy and Hard Limits

Moderation and Red Lines

The safety system operates on several levels. Before generation, the prompt is analyzed for potentially harmful content. During and after, the result is also checked. OpenAI uses both automated classifiers and manual review in complex cases.

So-called “hard limits” are separately highlighted – these are things the model will not do under any circumstances, regardless of the prompt's wording. Sexual content with children is one such case. OpenAI emphasizes that there are no “creative exceptions” or workarounds here.

Social Sharing of AI-Generated Video: A Platform Challenge

A Social Platform – A Separate Challenge

The Sora app is not just a generation tool; it is also a platform where users share their work. This means that in addition to standard safety issues, there are those typical of any social network: reposts, remixes, anonymous accounts, and viral spread.

OpenAI states it is developing its policies with this specificity in mind. In particular, this means stricter requirements for content that is publicly distributed within the platform compared to what a user creates for personal use.

Sora Safety Research Collaboration and External Oversight

Research and External Oversight

Before launching Sora, OpenAI engaged external security researchers – so-called “red teamers,” whose job is to intentionally look for ways to bypass protections. This is a standard practice in the industry, but it is especially important for video models, as the potential for misuse is much broader here than with text-based systems.

The company also collaborates with organizations dedicated to child protection and combating disinformation. This is not just a PR move – such partnerships provide feedback from people who see the real-world harm of such content in their work.

Open Questions and Ongoing Challenges in AI Video Safety

What Remains an Open Question

An honest look at the situation requires acknowledging that no protection system is perfect. Classifiers make mistakes. Malicious actors look for workarounds. The scale of the platform inevitably means that some unwanted content will slip through.

OpenAI is transparent about this. The company states directly that safety is a continuous process, not a state that can be achieved once and for all. This is an important disclaimer: it means that the rules and mechanisms will change as new threats and new ways of using them emerge.

For users, this is perhaps the main takeaway: Sora is a powerful tool with real limitations, and these limitations exist not to hinder creativity but to prevent the technology from becoming an instrument of harm. How well this balance can be maintained in practice, only time will tell.

Original Title: Creating with Sora Safely
Publication Date: Mar 23, 2026
OpenAI openai.com A U.S.-based company developing general-purpose AI models for text, code, and images.
Previous Article How to Safely Deploy AI Agents in Customer Support: The Notch Experience Next Article GitHub Taught Its Security Scanner to Understand Code Like a Human

Related Publications

You May Also Like

Explore Other Events

Events are only part of the bigger picture. These materials help you see more broadly: the context, the consequences, and the ideas behind the news.

Researchers have described a technique called 'promptomorphism': the generation of endless variations of malicious code by AI tools is making traditional defense methods increasingly ineffective.

Gen Digitalwww.gendigital.com Mar 21, 2026

Generative AI technology undeniably simplifies many aspects of life, yet questions regarding the safety and trustworthiness of generated content are emerging with increasing frequency.

Clova AIclova.ai Jan 23, 2026

From Source to Analysis

How This Text Was Created

This material is not a direct retelling of the original publication. First, the news item itself was selected as an event important for understanding AI development. Then a processing framework was set: what needs clarification, what context to add, and where to place emphasis. This allowed us to turn a single announcement or update into a coherent and meaningful analysis.

Neural Networks Involved in the Process

We openly show which models were used at different stages of processing. Each performed its own role — analyzing the source, rewriting, fact-checking, and visual interpretation. This approach maintains transparency and clearly demonstrates how technologies participated in creating the material.

1.
Claude Sonnet 4.6 Anthropic Analyzing the Original Publication and Writing the Text The neural network studies the original material and generates a coherent text

1. Analyzing the Original Publication and Writing the Text

The neural network studies the original material and generates a coherent text

Claude Sonnet 4.6 Anthropic
2.
Gemini 2.5 Pro Google DeepMind step.translate-en.title

2. step.translate-en.title

Gemini 2.5 Pro Google DeepMind
3.
Gemini 2.5 Flash Google DeepMind Text Review and Editing Correction of errors, inaccuracies, and ambiguous phrasing

3. Text Review and Editing

Correction of errors, inaccuracies, and ambiguous phrasing

Gemini 2.5 Flash Google DeepMind
4.
DeepSeek-V3.2 DeepSeek Preparing the Illustration Description Generating a textual prompt for the visual model

4. Preparing the Illustration Description

Generating a textual prompt for the visual model

DeepSeek-V3.2 DeepSeek
5.
FLUX.2 Pro Black Forest Labs Creating the Illustration Generating an image based on the prepared prompt

5. Creating the Illustration

Generating an image based on the prepared prompt

FLUX.2 Pro Black Forest Labs

Want to know about new
experiments first?

Subscribe to our Telegram channel — we share all the latest
and exciting updates from NeuraBooks.

Subscribe