OpenAI has outlined a safety framework for Sora that embeds protections into how AI-generated video content is created, shared, and managed.
The system introduces visible and invisible provenance signals, including C2PA metadata and watermarks, designed to ensure that generated media can be identified and traced.
The framework emphasises consent and control. Users can generate video content from images of real individuals only after confirming they have permission, while the ‘characters’ feature enables controlled use of personal likeness, with the ability to revoke access at any time.
Additional safeguards apply to content involving minors or young-looking individuals, with stricter moderation rules and enforced watermarking.
Safety mechanisms operate across the entire lifecycle of content. Generation is subject to layered filtering that assesses prompts and outputs for harmful material, including sexual content, self-harm promotion, and illegal activity.
These automated systems are complemented by human review and continuous testing to address emerging risks linked to increasingly realistic video and audio outputs.
The system also introduces protections specific to audio and user interaction. Generated speech is analysed for policy violations, and attempts to replicate the style of living artists or existing works are restricted.
Users of Sora retain control over their content through reporting tools, sharing settings, and the ability to remove material, reflecting a broader approach that aligns AI-generated media with safety, transparency, and accountability standards.
Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!
