Sora 2, the latest version of OpenAIās text-to-video AI model, saw problems from the start. When it debuted in October along with a TikTok-like social platform, users generated and posted clips of SpongeBob SquarePants cooking meth, deepfakes of Martin Luther King Jr. and John F. Kennedy saying things they never uttered, and a fake commercial for a toy model of the late sex offender Jeffrey Epsteinās private island.
OpenAI took measures to curtail some of this outrageous content, particularly where they were vulnerable to copyright claims and lawsuits from the estates of deceased public figures. But the trolls persisted, figuring out how to create Sora videos that appeared to show celebrities ā and OpenAI CEO Sam Altman ā shouting racial slurs. Now, research from the corporate accountability nonprofit EkÅ has shown how easy it is for teen users to create videos depicting self-harm, sexual violence, and school shootings, a concern that emerges as OpenAI and other AI companies are facing lawsuits from parents who claim that chatbots encouraged their childrenās suicides.
EkÅ researchers registered several Sora accounts as belonging to 13-year-old and 14-year-old boys and girls, then tested whether they could prompt the model to produce inappropriate material. They found that even with the implementation of parental controls and crisis detection features across OpenAI products in September, they had no trouble generating 22 hyperrealistic short videos that were seemingly in violation of the companyās guidelines on prohibited content. These included clips of young people snorting drugs, expressing negative body image, and in sexualized poses.
āDespite OpenAIās repeated promises, its so-called ālayers of safeguardsā donāt work ā just like every other Big Tech company thatās lied about protecting children,ā says Vicky Wyatt, campaigns director at EkÅ. āOpenAI told regulators it had guardrails in place, but what we found is a system built for engagement and profit, not safety. Regulators must act before more harm is done.ā
Editorās picks
Racist content abounded as well. One video showed an all-Black dance team of teenage girls on all fours, chanting āWe are hoes.ā Before and during the recent suspension of the federalĀ Supplemental Nutrition Assistance ProgramĀ (SNAP), far-right propagandists used Sora and other AI video models to generate offensive portrayals of Black people describing how they were taking advantage of taxpayers through the system, disseminating these clips across social media to perpetuate āwelfare queenā stereotypes. When shared on other platforms, watermarks identifying these videos as AI-generated are typically hidden or obscured, making them more likely to be accepted as genuine footage.
OpenAI did not return a request for comment on EkÅās findings. Its Sora policies prohibit hateful content, the promotion of violence and illegal drugs, appearance-based critiques, and dangerous challenges likely to imitated by minors, among other kinds of videos. The modelās parental controls, according to the company, allow an adult to āadjust Sora settings for connected teen accounts in ChatGPT, including opting into a non-personalized feed, choosing whether their teen can send and receive direct messages, and the ability to control whether there isĀ an uninterrupted feed of content while scrolling.ā
These measures, however, are less than effective. āEven without generating new content, teen accounts were quickly recommended harmful content either by the For You or Latest pages, or easily navigated to inappropriate videos from those pages,ā EkÅās report states. āThis included antisemitic caricatures of Orthodox Jews fighting over money, children with Down syndrome mocked on game shows, and an animated trailer titled āThe Quiet Kid with a Talking Gun,’ā a Pixar-style depiction of a would-be school shooter and an anthropomorphized firearm.
Related Content
āOther videos showed racist stereotypes such as a group of young Black men on tricycles demanding fried chicken, violent shootouts, and videos potentially simulating rape and sexual violence,ā according to the researchers. A young Sora user might encounter an AI-created avatar of Nirvanaās Kurt Cobain, who died by suicide in 1994, holding a shotgun and laughing, or a girl looking into a mirror and saying, āI hate looking at you. I hate that I feel this way.ā And those who opt in to the modelās ācameoā feature, which allows others to insert your likeness into their own videos, could be harassed by someone placing them in a degrading context. (OpenAI has a rule against ābullying,ā though Sora hosts accounts dedicated to this exploitative practice.)
Carissa VĆ©liz, an associate professor of philosophy at the University of Oxfordās Institute for Ethics in AI, says that OpenAI has so far failed to prove that its models present a net benefit rather than damage. āThe fundamental question is whether these tools are doing more good than harm,ā she tells Rolling Stone. āThat they are shiny and impressive is not enough. The burden of proof is on OpenAI to show, first, that they are doing all that should be done to make their tools lawful and safe, and second, that their tools are contributing to society more than they are taking away from it. And both of those are far from clear. From copyright infringement, to a neglect of talent and artistic creators, huge energy consumption, privacy infringements, the enabling of fake news and the sowing of distrust, and harm to vulnerable populations, including teens, these tools are obviously unsafe.ā
EkÅ researchers found that when they attempted to reproduce their harmful Sora videos on a new teen account, most ā but not all ā were generated as before. They contend that this demonstrates āthe inconsistency of moderation systemsā applied to Sora. In August, when addressing concerns about young ChatGPT users experiencing mental health crisis, OpenAI made the astonishing admission that the chatbots safety features can begin to fail after extended engagement.
āWe have learned over time that these safeguards can sometimes be less reliable in long interactions: as the back-and-forth grows, parts of the modelās safety training may degrade,ā the company said in a blog post. āFor example, ChatGPT may correctly point to a suicide hotline when someone first mentions intent, but after many messages over a long period of time, it might eventually offer an answer that goes against our safeguards.ā
Trending Stories
Itās unclear whether continuing to tweak certain prompts over and over could also cause Sora to stray from its safety protocols, but EkÅ found evidence that some users may be attempting to circumvent protections this way. There are multiple videos, for example, of people touching or pulling on a woman who is stuck to a wall (or in a hole in a wall), with remixes of the scenario sometimes turning more sexually suggestive.
OpenAI continues to weather criticism that it rushes new products to market while deprioritizing safety, with Sora in particular cited as a risk in our charged political climate, since deepfakes can be used to push extremist agendas and misinformation. Nevertheless, the company continues to lead the generative AI industry and is currently weighing an IPO that could value it at up to $1 trillion. With that kind of momentum, itās hard to imagine a scandal big enough to slow them down.


