AVisualHistoryofAIFilm&Video

From 2-second blurs to Lincoln Center premieres. My personal take on the evolution of generative AI video and film.

I included examples of my own work as a personal reference.

2019
to Present

2019-2021

BeforetheDawn

Before text-to-video models existed, a few pioneers hacked together ways to make AI generate moving images. Frame by frame, zoom by zoom, prompt by prompt. These experiments laid the aesthetic and technical foundation for everything that followed.

5 models1 films
February 2019Model Release
ThisPersonDoesNotExist.comPhillip Wang / NVIDIA StyleGAN

Uber engineer Phillip Wang launches a website that generates a new hyper-realistic human face on every page reload, powered by NVIDIA's StyleGAN. The site is visited 4.2 million times and becomes a cultural moment: the first time the general public confronts AI-generated imagery that is indistinguishable from photography. The concept showed that AI could generate photorealistic faces indistinguishable from real photographs, with no camera, no model, no studio. For creators, it planted a seed: if AI can imagine a face this convincing, what else can it imagine?

2019Film

Generating Photorealistic Video with AI

Jon Warlick

A wedding videographer from North Carolina creates what may be the first AI-generated video. Using NVIDIA's GauGAN, a tool that turns simple color-coded sketches into photorealistic landscape images, Warlick paints a landscape in GIMP, animates a camera movement through it in After Effects, exports every single frame, feeds each one individually through GauGAN, then reassembles the AI-transformed frames back into video. The result is flickering, dreamlike, and unmistakably alien. Three years before CogVideo, one person with patience and ingenuity found a way to make AI move.

Made with NVIDIA GauGAN

February 10, 2020Model Release
Video Enhance AI v1Topaz Labs

Topaz Labs releases Video Enhance AI, the first commercially available AI-powered video upscaling tool. Using deep learning models trained on thousands of video pairs, it can upscale footage up to 8K while adding genuine detail rather than just interpolating pixels. The tool quickly establishes Topaz Labs as the dominant player in AI video enhancement. While not a generative video tool, Video Enhance AI sets a critical foundation: when AI video models arrive years later with their low resolutions and noisy outputs, this is the tool creators will reach for to make that footage usable.

January 5, 2021Model Release
DALL-E AnnouncedOpenAI

OpenAI reveals DALL-E, a 12-billion-parameter version of GPT-3 trained to generate images from text descriptions. Type 'an armchair in the shape of an avocado' and get a photorealistic image back. It is not released to the public, but the concept is proven: language can direct visual creation. For creators, this is the 'holy grail' moment. The idea that you could describe what you want and have AI create it transitions from science fiction to demonstrated reality. Everything that follows in AI video builds on this foundation.

12B params
April 2021Model Release
VQGAN+CLIP Goes ViralKatherine Crowson & Ryan Murdock

Artist-mathematician Katherine Crowson (RiversHaveWings) and researcher Ryan Murdock combine two open-source models, VQGAN and CLIP, into a Google Colab notebook that anyone can run. Type a text prompt, get an AI-generated image. But the real breakthrough is animation: by applying zoom, rotation, and translation to each frame and feeding it back through the system, creators discover 'prompt travel', infinite zooms into AI-imagined worlds that shift and morph as the prompt changes. The aesthetic is surreal, psychedelic, unmistakably AI. Twitter accounts sharing VQGAN+CLIP art go viral. For the first time, a global community of artists is creating AI art together, sharing techniques, and pushing the medium forward. This is the birth of the AI art community that will later become the AI filmmaking community.

October 29, 2021Model Release
Disco Diffusion v1Somnai & Katherine Crowson

Somnai releases Disco Diffusion, the first tool to combine diffusion models with CLIP for text-guided image and animation generation. Unlike VQGAN+CLIP's iterative optimization, diffusion models generate images by gradually denoising random noise, producing higher quality results with more coherent structure. Somnai adds 2D animation techniques, and Adam Letts contributes 3D animation. The tool evolves rapidly through six major versions in eight months. Disco Diffusion becomes the training ground for the generation of AI artists who will later define the AI filmmaking movement. Riccardo Fusetti's AIFF 2023 Grand Prix winner 'Generation' will be made with this lineage of tools.

2022

TheDawn

Research labs crack text-to-video. The outputs are blurry, short, and uncanny. Nobody takes it seriously yet, but early creators had a lot of fun experimenting.

6 models1 films
May 2022Model Release
CogVideoTsinghua University / BAAI

The first open-source pretrained text-to-video model. 9.4 billion parameters generating 4-second clips at low resolution. Blurry, dreamlike, barely coherent... but it works. For the first time, anyone with a GPU could type words and get moving pictures back. The quality was terrible by any standard. But the concept was proven: machines could imagine motion from language. Everything that followed started here.

480x480 · 4s · 9.4B
September 2022Model Release
Make-A-VideoMeta AI

Meta's research team demonstrates that text-to-image models can be extended to generate video. The technique leverages what AI already learned about images and teaches it how things move. Never released to the public, but the demos circulate widely among researchers and signal that every major tech company is now racing toward AI video. The implications for creators are clear: if this works at scale, anyone will be able to direct scenes with words.

768x768 · ~5s
September 2022Model Release
Deforum Stable DiffusionDeforum Community

Weeks after Stable Diffusion goes public in August 2022, Deforum launches as a Colab notebook that turns text prompts into animated video. Building on the lineage of Disco Diffusion, VQGAN+CLIP, and PyTTI, Deforum adds Stable Diffusion's dramatically improved image quality to the animation workflow. The signature technique is 'prompt travel': schedule different prompts at different timestamps and watch the AI morph between them, creating dreamlike sequences where one world melts into another. Camera controls let creators zoom, pan, rotate, and fly through AI-generated landscapes. Quickly evolves into an Automatic1111 WebUI extension, making it accessible to anyone who can run Stable Diffusion locally. Deforum animations are projected at concerts worldwide, synced to music. The tool becomes the creative backbone of a generation of AI artists who will later move into AI filmmaking. Many of the aesthetic signatures of early AI video, the morphing transitions, the surreal prompt travels, the infinite zooms, originate here.

October 2022Model Release
Imagen VideoGoogle Brain

Google Brain enters the race with a cascade of video diffusion models producing 1280x768 video. The quality leap over CogVideo is dramatic, but like Meta's work, it stays locked in the research lab. What matters for creators: Google is proving that high-resolution AI video is possible. The gap between 'research demo' and 'tool you can use' is the only thing left to close.

1280x768 · ~5s
October 2022Model Release
Topaz Video AI v3Topaz Labs

Topaz Labs rebrands Video Enhance AI as Topaz Video AI with a major v3.0 rewrite featuring improved stabilization, faster processing, and new AI models for denoising, upscaling, and frame interpolation. The timing proves prescient: as generative AI video models begin producing short, low-resolution, noisy clips, Topaz Video AI becomes an essential part of the creator pipeline. Filmmakers submitting to early AI film festivals and contests like Gen:48 rely on it to upscale and clean up their AI-generated footage, turning rough model outputs into something presentable on the big screen. For a community pushing tools beyond their limits to make longer-form work, Topaz bridges the gap between what the models can output and what audiences expect to see.

October 2022Model Release
PhenakiGoogle Research

Google demonstrates something no one else has: AI-generated video that lasts over two minutes, driven by a sequence of text prompts that change over time. While other models max out at a few seconds, Phenaki shows that long-form AI video is theoretically possible. For filmmakers watching from the sidelines, this is the first hint that AI might one day generate not just clips, but scenes.

~2min
December 30, 2022Film

My Top AI Video Animations of 2022

enigmatic_e

It's hard finding longer-form AI generated video from 2022, but this curated compilation by enigmatic_e (one of the early pioneers) touches on some great examples from this era. A time capsule of the raw experimentation that what was possible before major generative AI video tools existed.

2023

TheCambrianExplosion

Consumer tools arrive. Runway ships Gen-1 and Gen-2. ModelScope goes viral. The first AI film contests launch. The community forms.

4 models11 films
January 2023Film

The Dog & The Boy

Netflix Japan / WIT Studio

Netflix Japan and WIT Studio release a short anime with AI-generated background art, making it the first time a major studio openly credits AI alongside human artists in a commercial release. The backgrounds were AI-generated, but the craft was human. It sparked a heated debate in the anime community about where the line should be drawn, but it also offered a glimpse of what was coming: AI as a production tool, not a replacement.

Made with AI background generation (unspecified)

February 2023Model Release
Runway Gen-1Runway

AI video leaves the research lab. Runway Gen-1 is the first commercially available AI video tool, letting creators upload real footage and transform it with AI — applying new styles and visual treatments guided by text or reference images. Not text-to-video yet, but the moment AI enters the creative workflow. Thousands of filmmakers, designers, and artists get their hands on it immediately.

Up to 896x512 · 4s
March 2023Model Release
Runway Gen-2Runway

A quick follow-up to Gen-1, Gen-2 improves upon the initial video-to-video capabilities with higher-fidelity video generation — and makes the leap to text-to-video. Type a prompt, get a 4-second video clip. The quality is rough, the clips are short, and the physics are wrong. None of that matters. What matters is that for the first time in history, anyone with a browser can generate original video from nothing but words. Runway Gen-2 creates the AI filmmaking community overnight. The experiments begin.

Up to 1536x896 · 4s
March 20, 2023Film

Digital Deity

Uncanny Harry

An early AI short film created in the raw days of consumer AI video, when the tools were limited and the aesthetic was defined by surreal artifacts and dreamlike distortion. Uncanny Harry crafted a piece that felt intentional rather than accidental — an early signal that creators were finding artistic voice inside the noise.

Made with Midjourney v4, voice.ai, iPhone apps for animation

March 21, 2023Film

When CZ meets Schwarzenegger

Framer

A viral AI comedy starring Binance CEO CZ and Arnold Schwarzenegger — stitched together from Midjourney stills, DALL-E 2 generations, and D-ID face animation, composited in After Effects. The Schwarzenegger voice is from voice actor @Govournator. Racked up 3.4 million views and became one of the earliest examples of AI-generated content breaking through to mainstream internet audiences.

Made with Midjourney, DALL-E 2, D-ID, Adobe After Effects

March 23, 2023Viral Moment

"Will Smith Eating Spaghetti"

Reddit user chaindrop / ModelScope

The video that made AI video generation mainstream. Horrifying, hilarious, impossible to look away from. Generated with ModelScope's text-to-video tool. Millions of views. The 'this changes everything' moment for the general public.

April 21, 2023Film

Property

Shane Willet

A comedic AI rap video where dogs and cats don't see Red or Blue — just territory. Made with Runway Gen-2 in its earliest days, Property showed that generative video didn't have to be serious or cinematic to land. The humor works because the AI aesthetic adds to the absurdity rather than fighting it.

Made with Runway Gen-2

May 26, 2023Film

NEMESIS

Shane Willet

After contacting their PenPals, they have now awakened the Nemesis. There is no turning back — you can't escape the mafia bass beats of the apocalypse. A music video that leans hard into the visual language of the Gen-2 era: warped textures, hallucinatory motion, and a dark energy that feels native to the tools. The aesthetic is unmistakably of its moment — defining what early Runway Gen-2 art looked and felt like.

Made with Runway Gen-2

July 8, 2023Film

THE OLD LADY

Abel Art

An early AI short film from Abel Art, crafted in the first months after consumer AI video tools became available. A haunting, atmospheric piece that demonstrated the emotional range possible even with the limited tools of mid-2023. Abel Art would go on to become a key contributor to community projects like the T2 Remake.

July 15, 2023Film

Lost Marbles

Dale Williams (The Reel Robot)

One of the earliest AI-generated anime shorts, created during the first wave of AI filmmaking in mid-2023. Made with the tools available at the time: Midjourney for imagery and early Runway Gen-2 for video generation. Williams would go on to become one of the most visible AI animators in the community, with over 7 million views and multiple festival awards. Lost Marbles is where that journey began.

Made with Midjourney, Runway Gen-2

July 31, 2023Film

SH*T!

Abel Art

A comedic AI short from Abel Art that leans into the absurdist potential of early generative video. Made weeks after THE OLD LADY, it shows a creator already experimenting with tone and genre — moving from atmospheric drama to irreverent humor. A reminder that the early AI film community wasn't all serious experimentation; some of the best work was just fun.

August 19, 2023Film

Mechaniverse

Shane Willet

An early attempt at building a visually consistent world using generative AI — mechanical creatures, clockwork landscapes, and impossible machines stitched into a cohesive short. Mechaniverse captures the distinct aesthetic of this period: dreamlike, slightly uncanny, rich with details, and pushing for narrative consistency when the tools barely supported it.

Made with Midjourney, Pika

August 31, 2023Film

NinjaPunk

Dave Clark

Set in the neon-lit streets of Los Angeles in 2065, a cybernetic ninja seeks vengeance after his wife's murder by the Yakuza, uncovering an underworld where ancient powers collide with futuristic warfare. Made with Midjourney and Runway Gen-2, this early short became a landmark: it later grew into a hybrid live-action/generative AI feature film developed at Promise Studios and showcased at the Cannes Film Market in May 2025, with stunt choreographers from John Wick. One of the most important early AI films.

Made with Midjourney, Runway Gen-2

November 2023Model Release
Stable Video DiffusionStability AI

Open-source video generation arrives. Stability AI releases models that turn a single image into a 2-4 second animated clip, and anyone can download and run them locally. For creators who want control over their tools, who want to fine-tune and customize, who want to run AI video without a subscription, this is the starting gun. The open-source AI video ecosystem begins here.

1024x576 · 2-4s
November 2023Festival / Contest

Gen:48 — 1st Edition

Runway Studios

2026

Dan Hammill & Jeff Wood

Grand Prix

The first 48-hour AI film contest. Over 1,000 participants given 48 hours to create a short film using Runway's tools. Dan Hammill and Jeff Wood win with '2026.' The AI filmmaking community has its first shared cultural moment.

November 2023Festival / Contest

AIFF 2023 — First Runway AI Film Festival

Runway

Generation

Riccardo Fusetti

Grand Prix

The first Runway AI Film Festival. Hundreds of submissions, screened in small New York theaters. 'Generation' by Riccardo Fusetti wins the Grand Prix — a philosophical journey through human experience created with Disco Diffusion layered over interpretive dance. AI film as art is born.

November 27, 2023Model Release
Magnific AIJavi López & Emilio Nicolás Riquelme.

Javi López & Emilio Nicolás Riquelme launch Magnific AI, the first popular 'creative' image upscaler — a tool that doesn't just add resolution but reimagines and hallucinates new details into images. With sliders for 'creativity,' 'HDR,' and 'resemblance,' creators can control how much the AI invents, turning rough Midjourney or Stable Diffusion outputs at the time, into images with stunning detail and texture. Over 30,000 people sign up within the first 24 hours. For the AI filmmaking community, where image-to-video workflows with models like Gen-2 depend highly on the quality of the input image, Magnific feels like magic. It fills in the details that early image models miss — skin pores, fabric texture, environmental detail — transforming good images into extraordinary ones. The impact is immediate and the approach is quickly emulated by competitors. Magnific is later acquired by Freepik in May 2024.

2024

TheGreatConvergence

Quality explodes. Big tech companies enter the race. Filmmakers get serious. The model wars begin.

11 models16 films
February 5, 2024Film

What if: The Takemura Files

FERA

A Cyberpunk 2077-inspired AI short that reimagines the game's world through generative video. FERA builds a moody, atmospheric narrative around the character Takemura, blending game aesthetics with AI-generated visuals. An early example of AI filmmaking drawing from gaming IP to create something that feels like a lost cutscene from a parallel universe.

February 2024Model Release
Sora PreviewOpenAI

OpenAI drops a preview that changes the conversation overnight. 60-second videos with coherent physics, consistent characters, and cinematic camera movements. Nothing publicly available comes close. The internet is split between awe and existential dread. For creators, the preview is both inspiring and frustrating: it proves AI video can be genuinely cinematic, but almost no one can actually use it. The gap between 'demo' and 'product' becomes the story.

Up to 1920x1080 · 60s
February 2024Film

Air Head

Shy Kids

One of the first films made with OpenAI's Sora. A 1:21 short about a person with a balloon for a head. Goes massively viral as part of Sora's public debut. Toronto collective Shy Kids becomes the face of early Sora filmmaking.

Made with Sora

February 2024Festival / Contest

Gen:48 — 2nd Edition

Runway Studios

Ellie

UNVEIL

Grand Prix

The second Gen:48. 1,600 signups. 'Ellie' by UNVEIL wins the Grand Prix. Filmmakers are given 48 hours and access to Runway's latest tools. The quality jump from the first edition is visible: the films feel more intentional, more polished, more like real filmmaking and less like experiments. The contest is proving that constraints (48 hours, AI tools) can fuel creativity rather than limit it.

March 3, 2024Film

The Illusion of Innocence

Rich Klein

(this is one of my animated short films)

My Gen:48 2nd Edition finalist film and my first real experience working with generative video. I worked with my daughter on the story using Runway Gen-2 for all the video generations. The film explores themes of innocence and perception through a fairy tale (with a twist).

Made with Runway Gen-2, Midjourney

March 6, 2024Film

Our T2 Remake

Sway Molina & Nem Perez / Storyblocker Studios

A feature-length AI parody remake of Terminator 2: Judgment Day, produced by over 50 AI artists and filmmakers. In this version, humanity must fight back against a world ruled by ChatGPT. Three months in the making, it premiered at the Landmark NuArt Theatre in Hollywood — one of the earliest community-driven, feature-length AI films to get a theatrical screening. A statement piece for the AI filmmaking community: proof that dozens of creators scattered across the internet could come together and ship a full-length movie.

May 2024Model Release
Veo 1Google DeepMind

Google officially enters the consumer AI video race at I/O 2024, announcing a model that generates 1080p video over a minute long. What this meant in practice: full HD video at lengths that start to feel like actual filmmaking, not just clips. Google's massive compute infrastructure means they can offer quality that smaller companies struggle to match.

1920x1080 · 60s+
May 2024Festival / Contest

AIFF 2024 — 2nd Runway AI Film Festival

Runway

Get Me Out

Daniel Antebi

Grand Prix

The second AIFF. 'Get Me Out' by Daniel Antebi wins Grand Prix — fast-paced, emotionally rich AI imagery flowing to music. The quality bar rises dramatically from year one. AI filmmaking proves it can move an audience.

June 12, 2024Model Release
Luma Dream MachineLuma Labs

Luma Labs launches Dream Machine and 25 million users sign up. The appeal is simplicity: type a prompt or upload an image, get video back fast. For many creators, Dream Machine is their first experience with AI video generation. It becomes the gateway that brings filmmakers, animators, and content creators into the AI video world. The sheer scale of adoption proves the demand is real.

June 2024Model Release
Kling 1.0Kuaishou

China's Kuaishou launches Kling, and within weeks it goes global. The model produces surprisingly good 1080p video that competes with Runway and Pika from day one. Kling represents something important: real competition in AI video. More models means more options, lower prices, and faster innovation. Kling will go on to iterate faster than any other model family in the space.

1920x1080
June 2024Model Release
Runway Gen-3 AlphaRunway

A major quality leap. Gen-3 Alpha produces 10-second clips with dramatically improved fidelity, motion coherence, and understanding of 3D space. For the first time, filmmakers start treating AI video as a real production tool rather than a novelty. The jump from Gen-2 to Gen-3 Alpha is the difference between 'interesting experiment' and 'I can actually use this in my work.' The era of AI filmmaking as craft begins.

10s
July 5, 2024Film

LE VOYAGEUR

Abel Art

A visually rich AI short from Abel Art that follows a traveler through dreamlike landscapes. By mid-2024, the tools had matured enough to match Abel Art's ambition — the result is a piece with real cinematic scope, atmospheric depth, and a visual confidence that reflects over a year of working with generative video.

2024Film

Borrowing Time

Dave Clark

Sci-fi short based on Dave Clark's father's true story of being Black in 1960s America. Made with Midjourney, Runway, and ElevenLabs. 110K+ views on X. Goes viral and leads to feature film offers from major production houses. Personal storytelling meets AI tools.

Made with Midjourney, Runway Gen-2, ElevenLabs

2024Film

The Last Artist

Dustin Hollywood

Billed as the world's first 100% AI-crafted feature-length film. Set in a future where AI has surpassed human creativity, the protagonist Tanaka stands as the last artist wielding a brush in a world dominated by digital perfection. Written and directed by Dustin Hollywood, founder of NAKID Magazine.

Made with Runway Gen-2

2024Film

What We Leave Behind

PJ Accetturo

A cinematic AI concept film depicting a Japanese village preparing for war. Made from scratch in just 3 hours using only $50 in Kling credits. PJ Accetturo showed that cinematic quality, speed, and low cost no longer had to be trade-offs in AI filmmaking.

Made with Kling

August 2024Model Release
Hailuo Video-01MiniMax

MiniMax quietly launches its first video model. 720p, 25fps, up to 6 seconds. The specs look modest on paper, but the motion quality and realism surprise everyone. Hailuo represents the emergence of a new player that will quickly become known for photorealistic output. The Cat Olympics viral moment a few months later will prove that Hailuo's physics simulation is genuinely convincing.

1280x720 · 6s
August 2024Viral Moment

Cat Olympics

Various creators

AI-generated videos of cats performing Olympic-style dives and gymnastics routines, created using the Hailuo model by MiniMax. Racks up millions of views and gets remixed endlessly across TikTok, YouTube, and Twitter/X. Proves AI video generation can be pure, joyful entertainment — no narrative ambition required.

Made with Hailuo / MiniMax

September 2024Film

Battalion

Dave Clark

A 5-minute short telling the story of the 320th Barrage Balloon Battalion, the only all-Black unit to storm the beach during D-Day. Created 100% with image-to-video and text-to-video generative AI using Runway and Luma. Goes viral with millions of views and leads to feature film offers from major production houses. Clark proves AI can tell stories that Hollywood won't.

Made with Runway, Luma Dream Machine

September 2024Film

The Yolky Universe: Concerto for Drips & Dreams

Sway Molina

A surreal, animated series by Sway Molina exploring beauty, horror, and the spaces between. The Yolky Universe was one of the most distinctive (and memorable to me) aesthetic signatures in early AI animation — weird, vivid, unmistakably handcrafted, even though it was AI-generated. As Molina put it: "The mind that's capable of creating beautiful worlds is the same mind that can create horrific ones for themselves. Always choose beauty and when in doubt — return to reality. It's not so bad here."

September 19, 2024Model Release
Kling 1.5Kuaishou

Kuaishou introduces Motion Brush, a feature that lets creators paint directional motion onto specific parts of the frame. Want a character's hand to wave while the background stays still? Draw the motion path. For creators, this is a breakthrough in creative control: you are no longer just typing prompts and hoping, you are directing the motion. Camera movement controls follow, making Kling feel less like a slot machine and more like a tool.

1920x1080
October 14, 2024Model Release
Adobe Firefly Video ModelAdobe

Adobe launches its Firefly Video model at Adobe MAX, entering the AI video race with a focus on commercial safety and professional workflows. Text-to-video with camera controls, image-to-video transformation, and Generative Extend for Premiere Pro that fills footage gaps and smooths transitions. Supports 100+ languages. The first AI video model explicitly designed for commercial use from day one, built on content that Adobe has the rights to use for training.

October 2024Festival / Contest

Gen:48 — 3rd Edition

Runway Studios

Beatrice

OLD NEW RARE

Grand Prix

The third Gen:48 introduces the IMAX Selection award, signaling Runway's ambition to bring AI films to the biggest screens. 'Beatrice' by OLD NEW RARE wins the Grand Prix. 'RUPU-KK: the Loop' takes People's Choice. Holden Boyles wins Best Craft for 'Honu,' Neural Viz wins Best Script for 'Welcome to Void's,' and Thomas Frenk earns the inaugural IMAX Selection for 'the brave little hen (or how a fox became a father).' The quality leap from Gen-2-era films to Gen-3-Alpha-era films is dramatic. These aren't experiments anymore. They're short films.

November 5, 2024Film

Saving the Sun

Rich Klein

(this is one of my animated short films)

One of my early animated shorts, made with Runway Gen-3. I wanted to explore themes of light, hope, and the lengths we'll go to preserve what matters most. A simple story, but one that felt meaningful to make.

Made with Runway Gen-3

November 12, 2024Film

One Last Wish

Edmond Yang, Simon Valentine & Alex Holm

A poignant short film about an elderly inmate's final wish, created in just 48 hours as part of the Cinema Synthetica Norway workshop. Script by acclaimed novelist Erlend Loe, directed by Simon Valentine, with AI visuals and animation by Edmond Yang. Won 1st Place in the Narrative Category at Project Odyssey 2025 and was featured in Forbes. A powerful example of how AI tools can accelerate traditional filmmaking workflows without replacing the human elements of storytelling, performance, and craft.

1st Place, Narrative Category — Project Odyssey 2025

Made with AI visuals (unspecified)

November 2024Viral Moment

Coca-Cola AI Holiday Ad

Coca-Cola / Silverside AI / Secret Level / Wild Card

Coca-Cola remakes its iconic holiday ad using Gen-3 Alpha, Luma Dream Machine, and Kling. The result is polarizing: praised by some for ambition, called soulless by others. Sparks a major debate about brands using AI to replace traditional production. The backlash proves the AI filmmaking community cares deeply about craft and authenticity.

Made with Runway Gen-3 Alpha, Luma Dream Machine, Kling

November 2024Festival / Contest

Gen:48 — 4th Edition

Runway Studios

Home

Ezra Li

Grand Prix

The fourth Gen:48 sees the competition mature. 'Home' by Ezra Li wins the Grand Prix. The roster of winners reads like an international AI filmmaking summit: Osvaldo Rivera Vazquez ('The King's Secret'), Faith Cho ('Clover'), Prasad Gori, Anurag Tiwari & Sagar Chogale ('Eyes on Me'), Nicholas Flandro ('No Pressure'), Mahmoud Salah & Karim Moftah ('I'm the Lamb'), Javier De La Chica & Guillermo Miranda ('!Dale caña!'), Matthew Berger, Kavin Kapoor & Ishan Modi ('How It's Made'), Alex Naghavi, Dan Taplin & Sara DeCou ('FEAST'), and Linus Mariscal Carulla, Max Ponce Pons & Dru Wu ('Aelita'). Recurring names across editions (Faith Cho, Javier De La Chica) show a professional community forming.

November 19, 2024Industry News
Promise AI Studio Launches

The first VC-backed studio built specifically for generative AI filmmaking. Co-founded by Dave Clark (Battalion, Borrowing Time), Fullscreen CEO George Strompolos, and former YouTube exec Jamie Byrne. Funded by Peter Chernin's The North Road Company and Andreessen Horowitz (a16z). Developing projects including 'NinjaPunk' and 'Extinction.' Building proprietary production software called MUSE. A signal that Hollywood money is betting on AI-native storytelling.

December 4, 2024Film

The Thin Man The Gun The Hotpot

Jim HuiHui

A 10-minute noir animation set in the neon-lit streets of Sichuan, China, following Xu Xia's struggle between conscience and survival. 242 shots, nearly three months of work, created largely by one person with AI. Midjourney generated 4,164 images, Kling produced 984 video clips, Pika added 693 more. Won Best Animated Short Gold at the Independent Shorts Awards. Featured at the AI Artist Festival and VACAT Awards. A landmark achievement in solo AI filmmaking: one creator, multiple AI tools orchestrated together, producing a film with the depth and craft of a studio production.

Best Animated Short Gold — Independent Shorts Awards

Made with Midjourney, Kling, Pika, Jimeng, PixVerse, Runway

December 2024Model Release
Veo 2Google DeepMind

Resolution jumps from 1080p to 4K, and the physics simulation takes a serious step forward. Released via VideoFX and later the Gemini app. Veo 2 becomes the model behind some of the most impressive creator work of the year, including Jason Zada's 'The Heist' (10M+ views) and Henry Daubrez's 'Kitsune.' Veo 2 is the first model where the output genuinely looks cinematic at feature-film resolution.

3840x2160 (4K)
December 2024Model Release
Sora Public LaunchOpenAI

Ten months after the preview that broke the internet, Sora finally launches for ChatGPT Plus and Pro users. The hype has cooled considerably. Competitors like Veo 2, Gen-3 Alpha, and Kling have closed the quality gap. The lesson is sobering: in AI, a year is a lifetime. The model that seemed impossibly ahead in February is merely competitive by December.

December 2024Film

The Heist

Jason Zada

Every shot generated text-to-video with Veo 2. A 1:56 bank heist and car chase. 1.1M views on the original X post, 10M+ across platforms. Proves that a single creator with taste and the right model can make something that looks like a studio production.

Made with Google Veo 2

December 19, 2024Model Release
Kling 1.6Kuaishou

Kuaishou claims a 195% overall improvement. Standard and Pro quality tiers give creators a choice between speed and fidelity. Better prompt adherence means the model actually does what you ask more often. More natural facial expressions and improved temporal consistency in sports and action scenes. For creators working on narrative projects, the reliability improvements matter more than any single feature.

720p / 1080p · 5-10s

2025

TheStudioEra

AI films go mainstream. IMAX screenings at Lincoln Center. Feature films get greenlit. Generative animation becomes its own art form.

23 models22 films
January 15, 2025Model Release
Luma Ray2Luma Labs

Built on an entirely new multi-modal architecture with 10x the compute of the original Dream Machine. Ultra-realistic details, coherent fast motion, and logical event sequences. For creators who built workflows around Dream Machine, Ray2 delivers a generational upgrade without changing the interface. The quality jump is dramatic enough that side-by-side comparisons circulate widely on X.

January 2025Film

Kitsune

Henry Daubrez

A tale of two lonely souls finding connection. Curated from 1,700+ AI-generated shots into a coherent narrative. Made with Veo 2. Named one of the 9 Best AI Animated Films of 2025 by Curious Refuge. A masterclass in AI film curation — the human taste is the tool.

Made with Google Veo 2

January 21, 2025Film

ALEX

Dale Williams (The Reel Robot)

An anime-style AI short from early 2025, showcasing Williams' signature style of character-driven narrative animation. Created during a highly productive period where Williams was collaborating with Promise AI on concept development and building his reputation as one of the leading AI anime filmmakers.

Made with Luma Labs, Midjourney

February 4, 2025Film

Rapture

BLVCKL!GHT

A feature-length AI-generated narrative, approximately one hour long, that transports viewers into another dimension through bizarre, surreal worldbuilding. Drawing from 1980s boundary-pushing shows like Ren and Stimpy and Pee-Wee's Playhouse, BLVCKL!GHT creates a sustained fever dream that proves AI filmmaking can work at feature length. Selected by Curious Refuge as one of the 9 Best AI Animated Films of 2025.

Made with Kling, Hedra, Luma, Runway, Hailuo, ElevenLabs

February 19, 2025Film

One Fire Left

Dale Williams (The Reel Robot)

An AI anime short depicting two surviving tribes fighting over Earth's last heat source as global temperatures drop to deadly levels. Won 1st Place in Curious Refuge's AI Animation Contest. Described as 'easily one of the best AI anime shorts out there.' Williams' work has amassed over 7 million views across social media and earned multiple festival awards.

1st Place — Curious Refuge AI Animation Contest

Made with Luma Labs

February 19, 2025Film

Gorbo's Swim Hole

BLVCKL!GHT

A surreal, darkly comedic piece set in an interdimensional water park where visitors encounter reality-bending experiences and may never truly leave. Part of BLVCKL!GHT's interconnected 'Route 47 Television' universe, an Adult Swim-inspired network of AI-generated shows. The concept grew into a recurring series with elaborate worldbuilding, its own website, and characters that cross between BLVCKL!GHT's other shows.

Made with Kling, Hedra, Luma, Runway, Hailuo, ElevenLabs

February 24, 2025Industry News
Promise AI Acquires Curious Refuge

Promise, the a16z-backed generative AI studio founded by Dave Clark, George Strompolos, and Jamie Byrne, acquires Curious Refuge, the leading AI filmmaking school. Founded in 2020 by Caleb and Shelby Ward, Curious Refuge has educated over 4,000 people from 171 countries through courses in AI filmmaking, animation, advertising, and documentary. The acquisition creates a direct pipeline from education to production: Curious Refuge creators can now work on Promise film and TV projects. For the AI filmmaking community, this is a signal that the ecosystem is maturing. The school that taught thousands of people to use AI tools is now part of the studio that will put those skills to work.

February 25, 2025Model Release
Wan 2.1Alibaba / Tongyi Lab

Alibaba open-sources a 14-billion-parameter video model under Apache 2.0, and it immediately tops the VBench leaderboard. For the first time, a state-of-the-art video model can run locally on consumer hardware with just 8GB of VRAM. The ComfyUI community adopts it overnight. For independent filmmakers who want to run their own pipeline without subscriptions or API costs, this changes everything. The democratization of AI video takes its biggest step forward.

1920x1080 · 5s · 14B / 1.3B
February 26, 2025Film

Cryptid Dating Game

BLVCKL!GHT

A twisted dating show parody where human contestants choose between grotesque cryptid bachelors like Mothman, a shapeshifting Skinwalker, and the recently divorced Jersey Devil. BLVCKL!GHT's breakout hit and most viral property, achieving tens of millions of views across TikTok and YouTube. The claymation-esque visual style blends horror, surrealism, and comedy. Proves that AI-generated serialized content can find massive audiences on short-form platforms.

Made with Kling, Hedra, Luma, Runway, Hailuo, ElevenLabs

February 27, 2025Industry News
House of David Uses AI for 72 Shots

Amazon's biblical epic House of David, created by Jon Erwin, premieres on Prime Video with 72 generative AI-assisted shots in Season 1 — one of the first major streaming series to openly integrate AI into its VFX pipeline. The standout is Episode 6's Goliath origin sequence: photoreal digital characters, armies of 10,000+, rain, smoke, and angel wing feathers — all driven by generative AI in collaboration with Vū Technologies. What would have taken four or five months of traditional post-production was completed in weeks. Erwin's approach: never generate from scratch, always augment existing show assets. The tools — Runway, Kling, Midjourney, Magnific, and Topaz alongside After Effects — handled physics simulations better than conventional VFX software at a fraction of the cost.

March 5, 2025Film

A Love Letter to LA

Paul Trillo, Paul Flores & Cuco

A poetic, music-driven tribute to Los Angeles directed by Paul Trillo — one of the most respected names bridging traditional filmmaking and AI tools. The film weaves generative imagery into a love letter that feels both deeply personal and unmistakably AI-native. Trillo's work consistently demonstrates that the best AI films come from directors who already know how to tell stories.

March 2025Model Release
Runway Gen-4Runway

The breakthrough feature is single-image character consistency: upload a reference photo and the model maintains that character's appearance across generated clips. For narrative filmmakers, this solves one of the biggest pain points in AI video. Before Gen-4, every generation was a roll of the dice on whether your character would look the same. Now you can build stories with recurring characters.

March 15, 2025Film

Sad Steve's Taxidermy and Ice Cream

BLVCKL!GHT

The tri-state area's only combination taxidermy studio and ice cream parlor. Sad Steve preserves deceased pets while his manic assistant Sprinkles serves up questionably-sourced frozen treats with flavors like 'Rocky Road Kill.' Part of BLVCKL!GHT's Route 47 Television universe, explicitly referencing characters from Gorbo's Swim Hole and The Mysterious Forest. The cross-series worldbuilding is a distinctive achievement in AI-generated content.

Made with Kling, Hedra, Luma, Runway, Hailuo, ElevenLabs

March 2025Film

LATE

Hal Watmough

First place winner at the Curious Refuge AI Animation Contest 2025. An unforgettable emotional narrative that showcases what generative animation can do when craft and feeling come first. Part of a growing wave of AI animation contests proving the medium has its own artistic identity.

1st Place — Curious Refuge AI Animation Contest 2025

April 15, 2025Model Release
Kling 2.0Kuaishou

A major quality leap with Multi-Elements Editor, 60+ style presets, and cinematic lighting controls. The creative control is the real story: you can now specify multiple elements in a scene and control how each one behaves. The style presets give you instant access to looks that previously required post-production. Kling is becoming a director's tool, not just a generator.

1920x1080 · 10s
April 18, 2025Film

Reflection feat

FERA

A music-driven AI short by FERA that pairs striking generative visuals with a hypnotic soundtrack. Fluid, reflective imagery merges with rhythm to create a piece that feels more like a music video from the future than a traditional short film — an example of AI filmmaking finding its own native format.

May 9, 2025Film

Zyphorax TV Tape 1

Uno Particular

A lo-fi, VHS-flavored fever dream that plays like a lost transmission from a public access channel in another dimension. Uno Particular leans into the glitchy, uncanny qualities of AI video rather than fighting them. The result feels like outsider art filtered through a broken CRT.

May 14, 2025Model Release
Wan 2.1-VACEAlibaba / Tongyi Lab

Alibaba combines video generation, reference-based generation, repainting, area modification, and spatio-temporal extension into a single open-source model. For creators using ComfyUI workflows, VACE turns Wan from a generator into a full production tool. Edit, extend, repaint, and modify shots without leaving the pipeline. One model that does the work of five separate tools.

14B / 1.3B
May 2025Model Release
Veo 3Google DeepMind

The landmark addition: native audio generation. For the first time, a video model generates synchronized dialogue, sound effects, and ambient noise alongside the visuals in a single pass. Lip-sync accuracy under 120ms. In practice, this collapses what was previously a multi-tool workflow into a single generation. No more generating video, then finding sound effects, then syncing dialogue. The model imagines the entire audiovisual scene at once.

4K+
May 2025Model Release
Kling 2.1Kuaishou

Introduces tiered quality modes: Standard for fast iteration, High Quality for final output, and a Master Edition for premium results. The tiered approach is practical: draft in Standard mode to find the right prompt and composition, then re-generate in Master for the final shot. Faster iteration without sacrificing quality when it counts.

May 25, 2025Viral Moment

Bigfoot Vlogs Go Viral

@bigfootvlogs / Various creators

Just one day after Google launches Veo 3 with native audio and dialogue, TikToker @bigfootvlogs posts an AI-generated video of Bigfoot casually vlogging in the woods: introducing himself, showing what he had for breakfast, taking a dip in a river, foraging for berries. It racks up over 6 million views in weeks. The format explodes across TikTok, X, and YouTube, with creators building entire character universes around talking cryptids: Yeti cooking shows, Mothman relationship advice, Loch Ness Monster nature documentaries. The trend proves two things: Veo 3's native audio generation is convincing enough to make fictional characters feel present and alive, and the AI filmmaking community has a sense of humor.

Made with Google Veo 3

June 1, 2025Film

THE CLEANER

Dave Clark

A thriller-style AI short featuring a hitman entering a room, crafted to feel like authentic Hollywood cinema. Clark challenged himself to create a coherent cinematic narrative using Google Veo 3 and Google Flow, achieving a level of shot-to-shot visual consistency rarely seen in AI film at the time. The film showed what Veo 3 looked like in the hands of an experienced filmmaker: remarkably realistic, Hollywood-grade footage with coherent lighting and staging across cuts.

Made with Google Veo 3, Google Flow

June 2025Film

Electric Pink

Henry Daubrez

A nostalgic coming-of-age memory set to motion. Made with Google Imagen 3 and Veo 2 via Flow. Premiered at the 2025 Tribeca Festival. Created for Google's I/O 2025 developer event. AI film reaches a prestigious traditional film festival.

Made with Google Imagen 3, Veo 2, Flow

June 2025Festival / Contest

AIFF 2025 — Lincoln Center with IMAX

Runway

Total Pixel Space

Jacob Adler

Grand Prix

The third Runway AI Film Festival, screened at Lincoln Center in partnership with IMAX. Over 6,000 submissions judged by Gaspar Noé, Harmony Korine, and Jane Rosenthal. 'Total Pixel Space' by Jacob Adler wins Grand Prix. 'JAILBIRD' by Andrew Salter takes Gold. 'ONE' by Ricardo Villavicencio & Edward Saatchi earns Silver. 'Distance Between Two Points Of Me' by Illya Dutsyk, Bogdan Nesvit & Anatolii Kasianov receives Honoree. The remaining finalists, 'More Tears Than Harm' (Herinarivo Rakotomanana), 'Fragments Of Nowhere' (Vallée Duhamel), 'RŌHKI - A Million Trillion Pathways' (Hachi and IO), '6000 Lies' (Simon Rieth), 'Editorial' (Riccardo Fusetti), and 'Emergence' (Maddie Hong), represent the global breadth of AI filmmaking. The jump from 300 submissions in 2023 to 6,000+ in 2025 tells the whole story.

June 16, 2025Model Release
Seedance 1.0ByteDance

ByteDance enters the AI video race with native bilingual support (Chinese and English), unified text-to-video and image-to-video, and multi-shot narrative generation with subject and style consistency. For creators, the multi-shot consistency is the standout feature: generate multiple shots of the same character in the same visual style, making it practical to build sequences that feel like they belong together.

June 18, 2025Model Release
Hailuo 02MiniMax

Native 1080p generation, 10-second clips, and state-of-the-art physics simulation. MiniMax's second model climbs to #2 on global benchmarks, surpassing Veo 3 on several metrics. Built with 3x more parameters and 4x more training data than Video-01. The leap in realism is immediately visible. Gymnastics, water physics, fabric movement: the things that used to look obviously fake now look convincing.

1920x1080 · 10s
July 15, 2025Model Release
Wan 2.2Alibaba / Tongyi Lab

The first open-source video model with Mixture-of-Experts architecture. 27 billion total parameters but only 14 billion active per step, cutting compute by 50%. Cinematic-grade quality with precise control over lighting, color tone, camera angle, composition, and focal length. For creators, Wan 2.2 becomes the community favorite: it balances quality, versatility, and cost better than anything else you can run locally.

1920x1080 · 27B total (14B active per step, MoE)
July 29, 2025Industry News
Cartoon Hero 1.0

Framer launches Cartoon Hero 1.0, a hands-on course community teaching workflows for creating AI-animated shorts. Students build cartoons from start to finish in multiple styles — sitcom, anime — learning tools and techniques for consistent character creation, multi-shot scenes, dialogue, lip-sync, and sound design. Impactful for demystifying AI animation workflows and proving that structured process, not just raw model output, is what makes AI cartoons feel professional.

August 2, 2025Model Release
Grok ImaginexAI

xAI releases Grok Imagine, offering text-to-image and image-to-video generation with sound. It produces short video clips from images — similar to Midjourney's video tool but lacking text-to-video capabilities. Elon Musk called it a "meme motherlode."

August 23, 2025Festival / Contest

Gen:48 — Aleph Edition

Runway Studios

Exactly How I Remember It

Holden Boyles

Grand Prix

The Aleph Edition, held on August 23, 2025, challenges filmmakers to use Runway's Gen-3 Alpha Turbo and Act-One tools to create short films in 48 hours. 'Exactly How I Remember It' by Holden Boyles wins the Grand Prix. Kartikye Kashyap takes People's Choice with '∞.' Ten winners are selected from a global pool of entrants. Recurring names from previous editions (YZA Voku, Javier De La Chica, Thomas Frenk) show a maturing professional community where the best creators keep pushing the boundaries with each new generation of tools.

August 26, 2025Model Release
Nano Banana (Gemini 2.5 Flash Image)Google DeepMind

Google's image generation model takes off under the codename 'Nano Banana.' Within two weeks, 23 million new users generate 500 million images. For AI filmmakers, this is the missing puzzle piece: a fast, high-quality image generator that maintains character consistency across outputs. Generate your character once, lock in the look, then feed the image to any video model as a starting frame. The image-to-video workflow that defines professional AI filmmaking finally has its source-of-truth tool.

Up to 4K
September 2, 2025Film

Steve Against the World

Rich Klein

(this is one of my animated short films)

This one’s different — my first real attempt at an animated comedy. Steve is a sarcastic, sophisticated cat with big opinions and a refined palate. He drinks lattes, avoids litter boxes, and judges other cats. I wanted to see if generative AI could sustain a recurring character across episodes. Turns out, it can.

Made with Veo 3, HeyGen Avatar IV

September 2025Model Release
Sora 2OpenAI

OpenAI's second attempt, released to select users in the US and Canada. Improved quality and longer outputs, but the competitive landscape has shifted dramatically since the original preview. Sora 2 is a solid tool that arrives in a crowded market. The window of Sora dominance has passed, but the model finds its audience among creators already embedded in the OpenAI ecosystem.

September 18, 2025Model Release
Luma Ray3Luma Labs

The world's first 'reasoning' video model. More than 2x the size of Ray2, it generates native 16-bit HDR video in ACES EXR format, bringing AI output directly into professional color grading pipelines. For creators working in professional post-production, this is a milestone: AI-generated footage that slots into existing studio workflows without quality compromises. The HDR output means AI video can finally sit alongside camera footage without looking out of place.

Native HDR (ACES EXR)
September 23, 2025Model Release
Kling 2.5 TurboKuaishou

Enhanced physics simulation, multi-step instruction handling, and causal relationships between elements in the scene. Nearly 30% lower pricing. What this meant for filmmakers: you can now describe complex sequences ('the ball rolls off the table, bounces twice, and the dog chases it') and the model understands cause and effect. Pricing drops make iteration affordable for independent filmmakers.

October 3, 2025Film

DRIFT SPLIT

The Butcher's Brain

A stylish, high-energy AI short blending drifting cars, split-screen editing, and kinetic visuals. The Butcher's Brain combines multiple video models — Veo, Kling, and Wan — with ElevenLabs for audio, showcasing the multi-model workflow that became standard practice in late 2025: pick the best tool for each shot rather than committing to a single pipeline.

Made with Veo, Kling, Wan, ElevenLabs

October 8, 2025Film

The BEST

Dale Williams (The Reel Robot)

An AI short that Williams described as 'probably the best thing I've ever done.' Created during the late 2025 period when more advanced AI video tools were available. Williams' work has been exhibited internationally and he has spoken at events including Pacific Film as a featured AI filmmaker.

Made with Midjourney, Luma, Kling

October 14, 2025Model Release
Veo 3.1Google DeepMind

Enhanced audio generation, improved realism, and powerful new editing features via Google Flow. Native 1080p HD with videos up to 60 seconds. The Flow integration is transformative: it creates a complete production environment around Veo, with editing, trimming, and refinement tools that turn raw generations into finished shots. The gap between 'AI generated clip' and 'production-ready shot' narrows significantly.

1920x1080 · 60s
October 2025Model Release
Hailuo 2.3MiniMax

Enhanced body movement, facial expressions, and physical realism. Introduces a Fast variant for batch creation at 50% lower cost. For creators working on volume projects like social content or storyboarding, the Fast variant means you can generate dozens of variations quickly and cheaply. The quality of the standard model continues to push toward photorealism, particularly in human movement and expression.

October 2025Industry News
All Heart Qualifies for Oscar with AI Animation

Oscar-winning filmmakers Michael Govier and Will McCormack (If Anything Happens I Love You) announce All Heart, a nine-minute animated short produced with Natasha Lyonne's AI studio Asteria. The film tells the story of a father visiting a young man with an extraordinary connection to his late daughter. The production combines hand-drawn artistry by illustrator Jimmy Thompson with AI animation: Asteria trained a custom model on Thompson's original art to generate additional production assets in his style, then integrated it into a traditional keyframing and in-betweening pipeline. The approach sidesteps the copyright debate entirely — the AI learns only from the creators' own work. The short qualifies for 2026 Academy Award consideration in Animated Short Film, making it one of the first AI-assisted films to enter the Oscar race.

October 16, 2025Film

Eyes On You

Rich Klein

(this is one of my animated short films)

My experiment in psychological horror with Veo 3 (Twighlight Zone style). A man visits a local theater for an ordinary movie night, only to discover the eyes on screen might not be acting. As he returns home, reality begins to blur.

Made with Veo 3

October 29, 2025Film

KYRA

MetaPuppet

A sci-fi AI short from MetaPuppet, one of the prolific creators in the AI filmmaking community. KYRA showcases the cinematic quality achievable with late-2025 generative video tools — polished visuals, coherent characters, and atmospheric worldbuilding from a solo creator.

November 20, 2025Model Release
Nano Banana Pro (Gemini 3 Pro Image)Google DeepMind

Best-in-class text rendering in images, multilingual text support, and integration with Google's world knowledge for accurate maps, diagrams, and infographics. Blends up to 14 images while maintaining consistency across 5 characters. For AI filmmakers, this becomes the 'virtual cinematographer': design your entire shot in a single image with precise character placement, lighting, and composition, then feed it to any video model.

Up to 4K
November 20, 2025Film

Hopeless Steve

Framer

A character who fails at everything and just can't turn his life around. Hopeless Steve became the breakout character from Framer's Cartoon Hero workflow — a lovably pathetic everyman whose animated misadventures proved that AI-generated cartoons could carry real comedic timing and emotional resonance. It also demonstrated that consistent animated characters with personality could be created using generative AI.

November 21, 2025Film

THE LAST PENCIL

The Butcher's Brain

A short film exploring what happens when the last traditional artist confronts a world that has moved on. The Butcher's Brain leans into Google's ecosystem — Veo 3 for video, Flow for editing, Gemini for planning — plus ElevenLabs for voice and Suno for music. A full AI production pipeline from a single creator, and a thematic mirror to the tool debate itself.

Made with Veo 3, Flow, Gemini, ElevenLabs, Suno

December 3, 2025Model Release
Kling 2.6Kuaishou

Native simultaneous audio-visual generation arrives at Kling: dialogue, singing, physics sound effects, environmental audio, and mixed mode. Following Veo 3's lead, Kling proves that audiovisual generation is becoming table stakes. The implications were immediate: fewer tools, faster iteration, and better synchronization between what you see and what you hear.

December 2025Model Release
Seedance 1.5 ProByteDance

Simultaneous audio-video generation with millisecond-precision lip-sync across 8+ languages and dialects. Cinema-grade camera controls including Hitchcock zoom, orbit, and tilt. For creators working in multilingual markets or dubbing, the language breadth is unmatched. The camera controls give directors vocabulary they understand: specify the exact camera movement you want, and the model executes it.

1920x1080 · 4-12s · 4.5B
December 2025Model Release
Runway Gen-4.5Runway

Runway's most advanced model delivers unprecedented visual fidelity and creative control. The culmination of the Gen-1 through Gen-4.5 evolution. For creators who have grown with Runway since Gen-1, this release represents how far the technology has come: from simple style transfer to photorealistic video generation with character consistency, all within three years.

December 16, 2025Model Release
Wan 2.6Alibaba / Tongyi Lab

China's first open-source reference-to-video model. Upload a character reference video with your appearance and voice, then generate new scenes starring yourself. This opened up open-source character consistency with audio sync. Anyone can now create long-form narratives with consistent characters using free, locally-run models. The combination of visual and vocal consistency opens the door to serialized AI content without subscription costs.

1080p · Up to 15s
December 18, 2025Model Release
Luma Ray3 ModifyLuma Labs

A specialized variant for hybrid AI workflows. Keyframe control lets you set start and end frames, Character Reference preservation maintains identity across edits, and video modification works while keeping consistency. For creators, this shifts Luma from pure generation to an editing paradigm: take existing footage, refine specific aspects, and maintain coherence throughout.

December 28, 2025Film

The Cleaner II - Swan Song

Dave Clark

The sequel to 'The Cleaner,' continuing the cinematic hitman narrative. Clark returns to the character and world he established in the first film, building a longer and more complex story. The fact that an AI filmmaker is building sequels and recurring characters marks a maturation of the medium from one-off experiments to serialized storytelling.

Made with Google Veo 3, Google Flow

September 2025Industry News
Google Flow Sessions

10 filmmakers given unlimited access to Google Flow plus mentorship. Henry Daubrez becomes Google Labs' first Resident Filmmaker. Google bets on individual creators, not just studios. A new model for AI film patronage.

2026

WhatComesNext

We're in uncharted territory. Feature films are in production. The tools generate sound and vision together. And the copyright question everyone's been avoiding is in the forefront.

7 models7 films
January 2026Industry News
Runway Widens AI Festival Beyond Film

Runway expands the scope of its annual AI Festival, adding categories beyond film. A signal that AI-generated media is spilling into games, interactive experiences, and new formats yet to be named.

January 26, 2026Model Release
Luma Ray3.14Luma Labs

Native 1080p generation at 4x faster speed and 3x lower cost. This eliminated the quality-speed-cost tradeoff that defined earlier models. You no longer have to choose between fast drafts and quality finals. Every generation is both fast and good, at a price point that makes iteration painless.

1920x1080
January 29, 2026Film

On This Day... 1776

Darren Aronofsky / Primordial Soup

The Oscar-nominated director of Requiem for a Dream and Black Swan releases an AI-generated animated series on TIME's YouTube, chronicling the American Revolution on its 250th anniversary. Made with Google DeepMind. SAG-AFTRA voice cast and original score. Each episode drops on the exact anniversary of the historical event it depicts. The reception is brutal: viewers and filmmakers call it 'AI slop.' Significant not for quality but for what it represents — a major Hollywood director founding an AI studio (Primordial Soup) and going all-in on AI animation.

Made with Google DeepMind (unspecified)

February 2, 2026Model Release
Grok Imagine 1.0xAI

xAI's "biggest leap yet" — Grok Imagine 1.0 unlocks 10-second text-to-video and image-to-video generation at 720p with dramatically improved audio. Instruction-following capabilities allow restyling scenes, adding/removing objects, and motion control. The API launched days earlier on January 28. By this point, Imagine had generated over 1.2 billion videos in a single month.

February 8, 2026Industry News
Grok Imagine Ad Contest

xAI runs an ad contest challenging creators to make commercials using only Grok Imagine 1.0, offering $1M for the grand prize. Over 4,000 submissions pour in. Andy Orsow wins $1M for "Understand the Universe. Faster." — a witty reimagining of Galileo's inquisition as a query to Grok. The winning ad airs during Super Bowl placements on X. Runner-up Dave Clark ($500K) and PJ Ace ($250K) round out the top three, with two $100K honorable mentions. A landmark moment for AI video contests in terms of prize money and mainstream visibility.

February 9, 2026Model Release
Seedance 2.0ByteDance

ByteDance's unified multimodal architecture accepts text, image, audio, and video as inputs. Up to 15-second clips at 2K resolution with extreme character consistency and audio-visual beat matching. For creators, the multimodal input is the breakthrough: feed the model a reference image, a voice clip, and a text description simultaneously, and it synthesizes all three into coherent video. Then copyright disputes from major studios pause the global rollout, raising questions about the training data behind the magic.

2K · 15s
February 2026Model Release
Kling 3.0Kuaishou

Unified omni-modal architecture spanning video, image, and beyond. From Kling 1.0 in June 2024 to 3.0 in February 2026, Kuaishou shipped 9 major versions in 20 months, the fastest model evolution in AI video history. For creators, Kling's trajectory proves that the best model today will be obsolete in months. The pace of improvement rewards creators who build adaptable workflows rather than betting on any single tool.

February 2026Film

Apex

The Dor Brothers

Marketed as a '$200M movie made in 24 hours with AI.' A nearly 3-minute sci-fi short that racks up 50M+ views across platforms. Featured on Joe Rogan. Made with Seedance 2.0. Divisive reception: praised for visuals, criticized for prioritizing spectacle over story. The Berlin-based brothers become the most-viewed AI filmmakers in the world.

Made with Seedance 2.0

February 2026Film

Overgrown

Henry Daubrez

A meditation on chronic resilience and the beauty of survival. Made with Google Flow and other tools during Daubrez's residency at Google Labs. In consideration for 2026 Tribeca Festival. One of the longest and most personal AI-generated films, pushing the format toward true short film territory.

Made with Google Flow, Veo

February 17, 2026Film

Feral - Concept Trailer

Dale Williams (The Reel Robot)

A concept trailer for a new project in Williams' signature anime-influenced style, showcasing the capabilities of early 2026 AI video tools. Williams has stated the industry is 'less than a year out from commercially viable AI films,' and concept trailers like this represent steps toward that goal.

Made with Seedance 2.0, Kling, Midjourney

February 19, 2026Model Release
Magnific Video UpscalerMagnific AI

Magnific brings the magic of its image upscaler to video. The same creative upscaling approach that transformed still images — adding details, enhancing texture, and adding fidelity beyond what the source contains — now works on moving pictures. For AI filmmakers who have relied on Magnific for their image-to-video pipelines since 2023, this is the natural next step: upscale the final output with the same creative intelligence that made the input frames shine. Precision mode was added in March 2026, focusing on preserving the original footage, while still allowing video enhancement options like 4K upscaling, sharpness, grain control, and FPS boost

February 26, 2026Model Release
Nano Banana 2 (Gemini 3.1 Flash Image)Google DeepMind

Pro-quality image generation at Flash speed, ranked #1 on the Artificial Analysis Image Arena at launch, at half the price of Nano Banana Pro. Maintains character consistency across 5 characters and 14 objects per workflow. For AI filmmakers, this completes the image-to-video production pipeline: generate consistent starting frames for every shot in your film, fast and cheap enough to iterate freely. Combined with the latest video models, like Seedance 2.0 or Kling 3.0, it enables cohesive visual storytelling.

Up to 4K
March 2026Film

Junkyard King ep.0

Henry Daubrez

An animated series pilot inspired by Arthurian legends in an 80s sci-fi setting. Made with Seedance 2.0, Google Flow, Gemini, Adobe Premiere/AE, ElevenLabs, and Suno. ~4 weeks of work, ~$1,000 in Seedance credits. Pioneered an 'omnireference' workflow. The future of serialized AI animation.

Made with Seedance 2.0, Google Flow, Gemini

March 2026Film

Costa Verde

Leo Cannone

A nostalgic masterpiece about childhood summers. Sweeps the AI International Film Festival March 2026: Best AI Film (Jury), Best Film (Audience), Most Surprising (Audience), Best Use of AI (Audience). Four awards for a single film. France-based Cannone proves AI film can be deeply emotional.

Best AI Film (Jury) + 3 Audience Awards — AI International Film Festival 2026

March 2026Film

Dragon Blue

Theoretically Media (Tim Simmons)

A hyper-stylized action thriller about vengeance and katanas. Every shot generated with Seedance 2.0. Created by Tim Simmons of Theoretically Media, one of the most prominent AI filmmaking voices on YouTube and a former Hollywood insider (Sony Animation). Simmons has been decoding AI tools for creators since 2022, and Dragon Blue shows what happens when deep technical knowledge meets cinematic ambition.

Made with Seedance 2.0

March 24, 2026Model Release
Sora ShutdownOpenAI

OpenAI announces that the Sora app and API are shutting down, 25 months after the preview that broke the internet. From the most hyped AI video model in history to discontinuation. For creators, the lesson is stark: platforms come and go, but your creative vision and storytelling skills transfer to whatever tool comes next.