OpenAI Releases Its First Text-To-Video Tool, Sora & Sparks Debate Over Ethics & Job Losses
It is a historic day for AI.
OpenAI just dropped the bomb on video generation.
Sora 🤯
That’s how everyone’s X-formerly-Twitter feeds look the last few days. Nearly every post on social media is about Sora, OpenAI’s first text-to-video tool that creates videos so impressive it’s scary.
According to OpenAI’s announcement, “Sora is capable of generating complex scenes featuring multiple characters, dynamic motion, and detailed backgrounds. Sora can generate videos up to a minute long while maintaining visual quality and adherence to the user’s prompt.”
The model is designed to understand how objects interact in the physical world and can accurately interpret props, generate expressive characters, and create videos based on still images, fill in missing frames, or extend existing videos.
“Sora is able to generate complex scenes with multiple characters, specific types of motion, and accurate details of the subject and background. The model understands not only what the user has asked for in the prompt, but also how those things exist in the physical world.”
And we must say that they really do look impressive.
According to OpenAI, though, the model has its weaknesses. As such, it may encounter difficulties accurately simulating the physics of intricate scenes and may fail to grasp particular cause-and-effect scenarios. For instance, while a person might be depicted taking a bite out of a cookie, the resulting image might not show the expected bite mark.
Spatial intricacies within prompts may confound the model, leading to occasional confusion between left and right orientations. Additionally, it might face challenges in providing precise descriptions of temporal sequences, such as tracking a predetermined camera trajectory over time.
“You’re literally hurting jobs with this.”
Despite the huge soar of excitement, it seems like doubt and concerns take centre stage anyway. Sora creates videos so realistic they might (will) actually take the jobs of lots of people, and those who now only consider careers in motion design and animation will have to think twice before applying to a related programme.
“The entire stock footage industry just died with this one tweet. RIP,” says one user on X. “It’s so over I’m going to lose my job,” comments another.
The whole comment section under the Sora launch announcement on X is filled with replies like this: “You scientists are so preoccupied with whether or not you can, you don’t stop to think if you should.” “OpenAI just can’t stop killing startups.” “I don’t think y’all realize how many artists you’re fucking over right now.” “If you see this @OpenAI please answer, in what way does this more good than bad? Like I am legit curious, yes I can see it’s cool but long term this will do so much damage is how I see it.” And this goes on and on.
Ethical data training concerns
After the launch, reporters and AI ethics advocate Ed Newton-Rex (former Stable AI executive and the founder of the nonprofit organisation Fairly Trained) rightfully enquired about how Sora has been trained: whose works scientists used to train Sora and whether the original creators of those works gave their explicit consent for their art to be fed to AI. There’s an assumption, though, that the model was trained on 3D simulations (in contrast to copyrighted material), as “all the generations look 3D rendered.”
As of writing, there’s no official OpenAI’s statement on the data sources.
“Absolutely zero ways in which this might be abused”
Another concern that has risen after the launch is how Sora might be used. Clearly, not all use cases will be innocent. In the age of deepfakes when no one is protected from being a victim of a harmful image or video, the concern is extremely valid.
According to OpenAI, “We’ll be taking several important safety steps ahead of making Sora available in OpenAI’s products. We are working with red teamers — domain experts in areas like misinformation, hateful content, and bias — who will be adversarially testing the model.
“We’re also building tools to help detect misleading content such as a detection classifier that can tell when a video was generated by Sora. We plan to include C2PA metadata in the future if we deploy the model in an OpenAI product.”