OpenAI’s new text-to-video AI, Sora, could transform how we create video content. Unlike previous, often nightmarish attempts by Google and Meta, Sora generates remarkably high-quality videos from simple text prompts or images. It can produce 1080p resolution with smooth motion – videos you might mistake for the real thing.
Early examples on OpenAI’s website reveal Sora’s impressive grasp of realistic visuals. It understands human proportions, complex lighting, and can even imitate classic film styles. While not flawless (subjects can appear unnaturally weightless and close inspection sometimes reveals AI quirks), Sora’s output consistently outperforms its predecessors.
Remarkably, Sora works with even short, open-ended descriptions. OpenAI even provides “bad” examples, like a man trying to run backward awkwardly on a treadmill, demonstrating transparency about the system’s current limitations.
We’re yet to see Sora’s ability to modify existing videos or insert new frames, which could prove revolutionary for editing and restoration tasks.
Before making Sora public, OpenAI is tackling critical concerns – potential misinformation, harmful content, and bias. The company said that they’re consulting experts and plan to include metadata for identifying AI-generated videos and granting limited access to a number of visual artists, designers, and filmmakers to gain feedback on how to advance the model to be most helpful for creative professionals.