- Runway unveils Gen-3 for AI-generated videos from text.
- Powerful controls over style, motion, filmmaking elements.
- Teases even more advanced future models coming.
Runway, a pioneer in generative AI tools for filmmakers, has taken the wraps off Gen-3 Alpha – its latest AI model capable of generating captivating video clips from mere text descriptions and still images.
Hollywood, we have liftoff
Boasting lightning-fast generation speeds and unparalleled fidelity, Gen-3 promises to elevate the art of video creation to new heights.
This cutting-edge model empowers creators to exercise fine-grained control over the structure, style, and motion of the videos it generates.
![](https://t9009141299.p.clickup-attachments.com/t9009141299/b331befc-bb4a-491c-9c9b-841d669b9384/HthXG8er5vRP4jD4mYLFW6-650-80.gif)
From expressive human characters to cinematic transitions and precise keyframing, Gen-3 Alpha speaks the language of filmmakers fluently.
The future on fast forward
While Gen-3 Alpha currently maxes out at 10-second clips and grapples with complex interactions, this is just the beginning.
![](https://t9009141299.p.clickup-attachments.com/t9009141299/9123252c-7df2-410b-9f68-974b2e07fcc0/image.png)
Runway co-founder Anastasis Germanidis teases a family of next-gen models trained on upgraded infrastructure, promising even greater capabilities on the horizon.