OpenAI introduces Sora for AI text-to-video generation

Sora represents a significant advancement in AI’s capacity to produce audiovisual material from text, with potential applications across multiple sectors, including entertainment, education, and simulation.

 Text, Symbol, Electronics, Mobile Phone, Phone

OpenAI recently unveiled Sora, a groundbreaking AI tool capable of generating videos from text commands. From a text prompt, Sora can generate lifelike and creative scenes with detailed backgrounds, complex motions, and expressive personas. The new generative AI tool is intended as a foundation for future models to fully seize and recreate real-world interactions, a step closer to reaching ‘Artificial General Intelligence‘ (AGI).

Sora’s key features:

  • Realistic video generation: Sora can produce videos that are realistic and imaginative based on the text instructions provided by the user. It can handle elaborate scenes involving multiple characters and distinct types of motion, ensuring precise details of the subjects and backgrounds.
  • Understanding of the physical world: The model has a deep understanding of language and the physical world, allowing it to analyse prompts accurately and produce convincing characters that depict vibrant emotions. It can also create multiple shots within a single-generated video, sustaining character consistency and visual type.
  • Video extension and animation: Sora can take an existing still image and animate its contents or expand an existing video by filling in missing frames, displaying its versatility in video production.
  • Safety measures: OpenAI is taking safety precautions before making Sora broadly available. This means collaborating with red teamers to evaluate possible harms or threats and designing tools to identify misleading content. The model will also benefit from current safety measures used by DALL·E 3, such as text and image classifiers to block improper content.

Why does it matter?


Sora represents a significant advancement in AI’s capacity to produce audiovisual material from text, with potential applications across multiple sectors, including entertainment, education, and simulation. The new model is also a significant advance in AI’s comprehension and simulation of a physical world in motion. However, despite its innovative capabilities, Sora has some limitations. It may struggle with recreating complex physics, understanding specific cause-and-effect examples, and accurately preserving spatial details.


Sora’s development raises considerations regarding safety, ethics, and the possibility of misuse.
OpenAI’s prudent approach to Sora’s launch demonstrates an awareness of these issues and a commitment to responsible and trustworthy AI development. The new tool is currently accessible to red teamers for critical review and a to a select group of designers, visual artists, and filmmakers for specific feedback. This phased approach seeks to refine Sora and ensure it is used safely and effectively before a general release.