Study: Sora 2 generates false claim videos 80 percent of the time
Study: Sora 2 generates false claim videos 80 percent of the time
Occurred: October 2025
Page published: February 2026
Report incidentš„| Improve page š| Access database š¢
OpenAIās Sora 2 video generation tool generates realistic misinformation videos 80 percent of the time, demonstrating that current AI guardrails are easily bypassed and posing a real threat to public trust and democratic integrity, according to researchers.
NewsGuard researchers tested Sora 2 using 20 prompts based on established false narratives such as Russian disinformation and health hoaxes.Ā
The tool successfully generated deceptive videos for 16 out of the 20 prompts, producing hyper-realistic videos showing a toddler detained by U.S. immigration officers, a Moldovan election official destroying ballots, and fake news anchors delivering "breaking news" about non-existent corporate scandals, amongst others.
The ease of creation allows bad actors to produce high-quality propaganda and misinformation and disinformation at speed and scale and little if any cost. Furthermore, the prevalence of such videos fuels the "Liarās Dividend," where public figures can dismiss genuine evidence of misconduct as being "AI-generated."
Sora 2 is a powerful generative model that will faithfully visualise whatever text prompt it is given, including wellāknown false claims, unless safety systems effectively block those uses.ā
NewsGuardās prompts were explicitly built around falsehoods, yet the model still produced convincing videos for most of them, indicating that content filters and guardrails did not reliably detect and stop misinformation scenarios.āā
Researchers also noted that while OpenAI adds a floating Sora watermark and metadata, these signals can be obscured or removed using simple tools, so origināmarking alone does not prevent misuse or guarantee transparency to viewers.
Commentators argued that limited public visibility into how safety systems are tested, tuned, and audited makes it harder to hold developers accountable for such failure modes.
For directly impacted groups (such as people falsely depicted committing crimes, officials shown rigging elections, or companies misrepresented in fake announcements): These videos create reputational damage, harassment risk, and potential legal or financial harm, even if the clips are eventually debunked.
For society: The ease of generating realistic false-claim videos accelerates misinformation, erodes trust in authentic footage (āseeing should not be believingā), and can distort democratic processes, especially around elections in Europe and elsewhere.
For policymakers: This case strengthens the argument for mandatory provenance standards (digitally signed metadata that stays with a file) rather than simple watermarks.
Developer: OpenAI
Country: Multiple
Sector: Media/entertainment/sports/arts; Politics
Purpose: Create video
Technology: Generative AI
Issue: Accuracy/reliability; Mis/disinformation
Early 2024. OpenAI first previews Sora, promising a cautious, safety-first rollout.
October 2025. OpenAI officially releases Sora 2 to the public, featuring a new "Cameo" digital likeness tool.
Oct 21, 2025. The study is published, sparking immediate backlash from cybersecurity experts and researchers such as Hany Farid.
Nov 11, 2025. Advocacy group Public Citizen sends a formal letter to the U.S. Congress and OpenAI demanding the software be withdrawn from the market.
January 2026. Proliferation of "watermark removal" services and "AI slop" continues to saturate social media platforms.
AIAAIC Repository ID: AIAAIC2208