OpenAI: Say Goodbye to Green Screens: This AI Creates Realistic Videos in Seconds
Sora is a an AI model developed by OpenAI team that can create realistic imaginative scenes from text prompts. It can create videos up to a minute long while maintaining visual quality and adherence to the user’s prompt.
Here’s an example from the OpenAI site:
Few sample videos generated by OpenAI Sora
How does OpenAI Sora work ?
Sora is a diffusion model, it generates a video by starting off with one that looks like static noise and gradually transforms it by removing the noise over many steps. The videos can be up to 60 seconds long.
Sora builds on past research in DALL·E and GPT models. It uses the recaptioning technique from DALL·E 3, which involves generating highly descriptive captions for the visual training data.
Limitations of OpenAI Sora
While Sora excels at bringing text to life with video, it’s still under development and has areas to improve. In intricate scenes, it might not perfectly capture real-world physics, leading to occasional glitches like a cookie bite magically disappearing.
Similarly, it’s learning to navigate spatial relationships and might sometimes mix up left and right. Additionally, precise descriptions of events unfolding over time, like a specific camera movement, can pose a challenge.
What are the risks of OpenAI Sora ?
- Generation of harmful content
- Misinformation
- Biases and stereotypes
How will OpenAI ensure safety of the Product usage ?
OpenAI is building tools to help detect misleading content such as a detection classifier that can tell when a video was generated by Sora.
How Can I Access Sora?
Sora currently available to only “red teams“(experts researchers).They will try to assess critical areas for harms or risks. We are also granting access to a number of visual artists, designers, and filmmakers to gain feedback on how to advance the model to be most helpful for creative professionals.