Runway Video-to-Video example output
NEW

Runway Video-to-Video

Transform existing footage with AI. Style transfer, scene reimagining, material override.

Upload a video and describe what you want it to become. Runway Video-to-Video uses the Gen-4 Aleph architecture to restyle, retexture, and reimagine footage while preserving the original motion, timing, and composition. The source video provides the structure. Your prompt provides the new look. The result is a transformed clip that moves identically to the original but exists in a completely different visual world.

Example outputs

Runway Video-to-Video example 1

Transform this office walkthrough into a warm mid-century modern interior with walnut panelling, brass fixtures, and soft pendant lighting

Runway Video-to-Video example 2

Restyle this product video as a hand-drawn pencil animation with visible cross-hatching and paper texture, keep the rotation timing

Runway Video-to-Video example 3

Convert this urban street footage into a snowy winter scene with frost on windows, breath visible, warm shop lighting spilling onto wet pavement

Runway Video-to-Video example 4

Transform this fashion runway clip into a painterly oil painting style with visible brushstrokes, warm gallery lighting, impressionist palette

Runway Video-to-Video example 5

Restyle this car driving footage as a cyberpunk night scene with neon reflections on wet roads, holographic billboards, purple and teal colour grade

Runway Video-to-Video example 6

Convert this garden walkthrough into an autumn version with golden and red foliage, long shadows, warm low-angle sunlight, fallen leaves on paths

How it works

01

Describe your scene

Type a detailed prompt describing the video you want, or upload a reference image as a starting frame.

02

Choose your settings

Pick your resolution and duration. See the credit cost before you generate.

03

Generate your video

Your video is ready in 1-3 minutes. Download, iterate, or extend the sequence.

Ready to create with Runway Video-to-Video?

Jump into the Studio and start generating. Plans from £10/month.

Choose a Plan

Same motion, completely different world.

Video-to-Video solves a problem that text-to-video cannot. When you already have footage with the right motion, timing, and composition but need it in a different visual style, generating from scratch means losing all of that choreography. V2V preserves it. Upload a walkthrough of a space and restyle it from concrete brutalism to warm timber. Take a product video and shift the lighting from cold studio to warm editorial. Convert a live-action reference into a stylised animation. The motion stays. The world changes.

The Gen-4 Aleph architecture analyses your source video frame by frame, extracting motion vectors, depth relationships, and compositional structure. It then regenerates every frame according to your text prompt while locking to the original timing. Camera movements track identically. Subject positions hold frame for frame. The result is not a filter applied on top of footage. It is a full regeneration that happens to follow the same choreography as your source material.

For design professionals, V2V opens workflows that were previously impossible without a full reshoot. An architect can show the same walkthrough in three material palettes. An interior designer can present a room tour in summer and winter lighting. A product designer can show the same prototype video in brushed aluminium, matte ceramic, and painted wood. Each version takes minutes and a text prompt instead of a new production day.

Style transfer with structure

V2V is not a colour filter. It regenerates every frame from the ground up, applying new materials, lighting, and environmental detail while locking to the motion and composition of your source video. The result holds up at full resolution because it is generated, not processed.

Material and environment override

Describe a new material palette and V2V applies it to every surface in the video. Concrete becomes timber. Glass becomes frosted. Daylight becomes golden hour. The depth relationships and lighting physics update to match the new materials, producing results that look physically coherent rather than painted on.

Frequently asked

Questions about Runway Video-to-Video.

Upload a video and describe what you want it to become. Runway Video-to-Video uses the Gen-4 Aleph architecture to restyle, retexture, and reimagine footage while preserving the original motion, timing, and composition. The source video provides the structure. Your prompt provides the new look. The result is a transformed clip that moves identically to the original but exists in a completely different visual world.
Built differently

Why Stensyl?

A small indie studio building creative tools the way they should be built. No VC theatre, no funnel games, no faceless support.