OpenAI’s Sora: What Happened, and Why Visla Isn’t Worried

Quick Answer

OpenAI shut down its Sora AI video generation app on March 24, 2026, just six months after launch, citing a need to focus compute resources on enterprise productivity tools. The real story is more complicated: Sora was burning an estimated $15 million per day in compute costs while generating only $2.1 million in total lifetime revenue, all while drowning in copyright lawsuits and deepfake controversies that killed a $1 billion Disney partnership before it ever paid out. Sora’s collapse raises broader questions about OpenAI’s competitive moat, the viability of consumer AI video, and whether the technology itself is ready for the ambitions people had for it.

Video generated using Visla

What OpenAI Said

OpenAI’s official explanation for shutting down Sora was tidy and strategic-sounding. Fidji Simo, OpenAI’s CEO of Applications, told employees the company needed to stop chasing what she called “side quests” and concentrate on high-impact enterprise tools. Sam Altman framed it as a resource allocation decision, focused on capital raising, data center buildout, and compute supply chains. The Sora research team was officially reassigned to “world simulation research” to support robotics development.

What Actually Happened With OpenAI’s Sora

The economics of Sora were simply unsustainable. According to estimates from Cantor Fitzgerald analyst Deepak Mathivanan reported by CNBC, each 10-second Sora clip cost OpenAI roughly $1.30 in compute, which adds up fast when you’re running a consumer app with millions of users. Against annualized infrastructure costs that analysts estimated in the billions, Sora generated just $2.1 million in total lifetime revenue from in-app purchases, per data from Appfigures. Sora’s own team lead, Bill Peebles, acknowledged publicly in October 2025 that “the economics are currently completely unsustainable.”

Beyond the money, Sora had a serious legal problem that OpenAI was never going to solve cleanly. Within days of Sora 2’s launch in late 2025, the platform was flooded with AI-generated videos of copyrighted characters: SpongeBob, Pikachu, Mario, South Park characters, and plenty more. Users generated deepfakes of celebrities and public figures. OpenAI initially tried an opt-out policy for rightsholders, which the Motion Picture Association blasted publicly, with Chairman Charles Rivkin calling for “immediate and decisive action.” Stanford Law professor Mark Lemley told CNBC that OpenAI was opening itself up to substantial copyright litigation. Within 72 hours, OpenAI reversed to an opt-in policy, but the reputational and legal damage was done. Multiple lawsuits followed, and the $1 billion Disney partnership, which would’ve covered 200+ characters across Disney, Marvel, Pixar, and Star Wars, collapsed entirely. Disney reportedly received 30 minutes’ notice before the public announcement.

And then there were the users themselves. Downloads dropped from 3.33 million in November 2025 to 1.13 million by February 2026, according to TechCrunch’s reporting on the shutdown. Thirty-day retention fell to single digits. People downloaded it, played around, made some weird clips, and left. There wasn’t a compelling, repeatable reason to stay.

The economics of OpenAI Sora
Unsustainable economics

Each Sora clip cost OpenAI ~$1.30 in compute

Cantor Fitzgerald analyst Deepak Mathivanan estimated each 10-second video generation cost roughly $1.30 — while Sora earned just $2.1M in total lifetime revenue.

Cost per 10s clip
$1.30
Total lifetime revenue
$2.1M
Infra costs (annual est.)
$Billions
Revenue source
In-app purchases
Clip length 10 sec
Number of clips generated 1,000
Estimated compute cost
$1,300
That’s 0.06% of Sora’s total lifetime revenue — burned on just 1,000 clips.
Cost vs. $2.1M lifetime revenue 0.06%

What This Means for OpenAI and the AI Landscape

Sora’s failure isn’t just a product story. It’s a signal about OpenAI’s broader position in the market, which is looking shakier than the company’s $840 billion valuation might suggest.

JPMorgan, in a rare analysis of a private company, described OpenAI’s competitive moat as “increasingly fragile.” Analyst Benedict Evans, formerly of Andreessen Horowitz, put it more bluntly in a February 2026 piece, arguing that foundation model labs have “no moat or defensibility except access to capital” and noting that ChatGPT engagement is “mile-wide, inch-deep,” with 80% of weekly active users sending fewer than 1,000 messages across all of 2025. Meanwhile, according to Menlo Ventures’ 2025 enterprise survey, OpenAI’s API market share fell from 50% in 2023 to just 25% by mid-2025, with Anthropic now leading in enterprise at 40%. In the coding category specifically, Anthropic commands 54% versus OpenAI’s 21%. ChatGPT’s consumer market share has slipped from roughly 87% to 68% year-over-year per Similarweb, with Google Gemini surging to fill the gap.

The AI video market itself remains real but small, estimated at around $700-$800 million in 2025, and the “generate anything, for anyone, for nearly free” consumer model has proven economically catastrophic for every company that’s tried it at scale.

Why Visla Is Built Differently

This is where we want to be direct: Visla is, in part, an AI video platform. And yet essentially none of what killed Sora applies to us. Here’s why.

We solved the consistency problem

One of Sora’s most frustrating technical failures was its inability to keep characters, objects, and environments consistent across multiple shots. Professional editors testing the tool reported regenerating the same 10-second clip dozens of times trying to get two consecutive scenes where the main character looked like the same person. Visla’s AI Director Mode was specifically built to address this. You choose your characters, objects, and environments before generation begins, and the platform keeps them locked across every scene in your storyboard. Your visuals don’t “change actors” mid-video. For the businesses and creators who use Visla to make product demos, training videos, explainers, and internal communications, this consistency isn’t a nice-to-have feature. It’s the whole point.

We use Google Veo’s API, and that’s a strategic advantage

Visla generates AI video clips via Google Veo’s API, which means we’re not carrying the massive, unsustainable infrastructure cost that sunk Sora. We benefit from Google’s enormous investment in making Veo better and cheaper over time, and Google has both the financial resources and the competitive incentive to keep doing exactly that. Google Veo is the most prominent surviving platform in AI video generation right now, and we’re building on it rather than competing with it.

AI video generation is one tool in a much larger toolkit

This might be the most important distinction of all. Sora was a single-purpose AI video generator. You typed a prompt, you got a clip, and that was it. Visla is a complete, start-to-finish video production platform. You can use AI-generated video clips if that’s what a project calls for. You can also:

  • Record your screen, webcam, or phone camera and build a video around that real footage
  • Use professionally produced B-roll from Storyblocks (2 million+ clips on Pro, 16+ million on Business and above via Getty Images and Storyblocks combined)
  • Upload your own video assets into Private Stock, where Visla’s AI auto-tags your footage and makes it available for any future project
  • Mix and match all of the above in the same video

You don’t have to touch AI video generation at all if your project doesn’t need it. That flexibility is exactly what a real production workflow requires.

We’re not in the business of generating gimmicky AI videos

Sora’s consumer model, by design, encouraged rapid, low-stakes clip generation with no particular purpose or structure. Visla’s model is different. We’re built around producing cohesive, longer-form videos that actually tell stories, communicate information, and serve a real purpose for a real audience. Whether that’s an onboarding video for new employees, a product walkthrough for customers, a training module for a specific workflow, or a piece of thought leadership content for LinkedIn, the goal is always a finished video that works. That’s not a moral stance. It’s just what our users actually need from us.

FAQ

Will OpenAI ever return to AI video generation?

OpenAI hasn’t ruled out re-entering the AI video space entirely, and Sam Altman has hinted that video generation could resurface as a feature within ChatGPT rather than as a standalone product, which would let the company avoid the unsustainable economics of a dedicated consumer app. That framing, video as one capability among many rather than the whole product, is actually closer to how Visla approaches it, and it suggests the broader industry is converging on the same lesson. Whether OpenAI makes a serious second attempt will likely depend on how much compute costs fall over the next two to three years and whether a clearer enterprise use case emerges that justifies the infrastructure investment.

Who benefits most from Sora’s shutdown?

The clearest short-term winners are Google Veo, Runway, and Kling, which now absorb users and enterprise attention that had been split with Sora, and all three have been moving aggressively to capitalize on the opening with new model releases and expanded API access. Longer-term, the real beneficiaries may be purpose-built platforms that combine AI video generation with a full production workflow, since Sora’s failure has essentially validated that model over the “just a generator” approach. Enterprise software buyers who were hesitant to commit to any AI video vendor now have stronger signals about which parts of the market are viable and which aren’t, which should actually accelerate purchasing decisions rather than slow them down.

Does the Sora shutdown signal that AI video isn’t ready for enterprise use?

It’s really the opposite: Sora’s failure was specifically a consumer product failure, not an indictment of AI video in enterprise contexts, where companies like Synthesia and Hedra are growing steadily by solving defined, high-value problems like employee training, product demos, and scalable internal communications. The core technical limitations of AI video, including consistency across shots and physics accuracy, matter far less when you’re building a structured corporate explainer than when you’re trying to generate cinematic fiction from a freeform prompt. Enterprise buyers don’t need AI video to be magic; they need it to save time, reduce production costs, and produce output that’s good enough to do its job, and today’s tools, used correctly, can already clear that bar.

May Horiuchi
Content Specialist at Visla

May is a Content Specialist and AI Expert for Visla. She is an in-house expert on anything Visla and loves testing out different AI tools to figure out which ones are actually helpful and useful for content creators, businesses, and organizations.


Enjoyed this article? Share your experience with Visla on G2: leave a review here

Join our thousands of subscribers.

Subscribe to our weekly newsletters for curated blog posts and exclusive feature highlights. Stay informed with the latest updates to supercharge your video production process.