Runway, one of several AI startups developing video-generating tech, todayannounced an API to allow devs and organizations to build the company’s generative AI models into third-party platforms, apps, and services.
Currently in limited access (there’s awaitlist), the Runway API only offers a single model to choose from — Gen-3 Alpha Turbo, a faster but less capable version of Runway’s flagship, Gen-3 Alpha — and two plans, Build (which is aimed at individuals and teams) and Enterprise. Base pricing is one cent percredit (one second of video costs five credits), and Runway says that “trusted strategic partners” including marketing group Omnicom are already using the API.
The Runway API also comes with unusual disclosure requirements. Any interfaces using the API must “prominently display” a “Powered by Runway” banner linking to Runway’s website, the company writes in a blog post, to “[help] users understand the technology behind [applications] while adhering to our usage terms.”
Runway, which isbacked by investors including Salesforce, Google, and Nvidia and was last valued at $1.5 billion, faces stiff competition in the video generation space, including from OpenAI,Google, andAdobe. OpenAI is expected to release its video-generation model,Sora, in some form this fall, while startups likeLuma Labs continue to refine their technologies.

Case in point, in what seems unlikely to be coincidental timing, Luma todaylaunched its API for video generation, which doesn’t have a waitlist and offers features beyond Runway’s, including the ability to “control” the virtual camera in AI-generated scenes.
With the preliminary launch of the Runway API, Runway becomes one of the first AI vendors to offer a video-generation model through an API. But while the API might help the company along its path to profitability (or at least recouping the high costs of training and running models), it won’t resolve the lingering legal questions around those models and generative AI tech more broadly.
Runway’s video-generating models, like all video-generating models, were trained on a vast number of examples of videos to “learn” the patterns in these videos to generate new footage. Where did the training data come from? Runway refuses to say, like many vendors these days — partly out of fear of sacrificing competitive advantage.
Join the Disrupt 2026 Waitlist
Add yourself to the Disrupt 2026 waitlist to be first in line when Early Bird tickets drop. Past Disrupts have brought Google Cloud, Netflix, Microsoft, Box, Phia, a16z, ElevenLabs, Wayve, Hugging Face, Elad Gil, and Vinod Khosla to the stages — part of 250+ industry leaders driving 200+ sessions built to fuel your growth and sharpen your edge. Plus, meet the hundreds of startups innovating across every sector.
Join the Disrupt 2026 Waitlist
Add yourself to the Disrupt 2026 waitlist to be first in line when Early Bird tickets drop. Past Disrupts have brought Google Cloud, Netflix, Microsoft, Box, Phia, a16z, ElevenLabs, Wayve, Hugging Face, Elad Gil, and Vinod Khosla to the stages — part of 250+ industry leaders driving 200+ sessions built to fuel your growth and sharpen your edge. Plus, meet the hundreds of startups innovating across every sector.
But training details are also a potential source of IP-related lawsuits if Runway trained on copyrighted data without permission. There’s evidence that it did, in fact — areport from 404 Media in July exposed a Runway spreadsheet of training data that included links to YouTube channels belonging to Netflix, Disney, Rockstar Games, and creators like Linus Tech Tips and MKBHD.
It’s unclear whether Runway ultimately ended up sourcing any of the videos in the spreadsheet to train its video models. In an interview with TechCrunch in June, Runway co-founder Anastasis Germanidis would only say that the company uses “curated, internal datasets” for model training. But if it did, it wouldn’t be the only AI vendor playing fast and loose with copyright laws.
Earlier this year, OpenAI CTO Mira Murati didn’t outright deny that Sora was trained on YouTube content. And Nvidiareportedly used YouTube videos to train an internal video-generating model called Cosmos.
Many generative AI vendors believe that the doctrine known asfair use provides legal cover — andthey’re asserting this in courtand inpublic statements. Others are less inclined to take chances, and/or they view a more “ethical” approach to model training as a selling point for their services. To develop its video-generating Firefly models, Adobe is said to beoffering artists payments in exchange for clips, for example.
In its terms of service, Luma says that it’ll agree to defend and indemnify API business customers for damages arising out of IP violation claims. Other vendors, including OpenAI, offer similar indemnification policies; Runway does not, although it said last December that it would partner with stock media library Getty to develop more “commercially safe” versions of its products.
However the lawsuits pertaining to the legality of training on copyright content shake out, one thing’s becoming clear: Generative AI video tools threaten to upend the film and TV industry as we know it.
A 2024 study commissioned by the Animation Guild, a union representing Hollywood animators and cartoonists, found that 75% of film production companies that have adopted AI have reduced, consolidated, or eliminated jobs after incorporating the tech. The study also estimates that by 2026, more than 100,000 U.S. entertainment jobs will be disrupted by generative AI.
Updated 9/16 at 11:18 Pacific: Added info about Luma’s API launch.
Topics

StrictlyVC concludes its 2025 series with an exclusive event featuring insights from leading VCs and builders such as Pat Gelsinger, Mina Fahmi, and more. Plus, opportunities to forge meaningful connections.

