Workflow Guides

Seedance 2.0 Character Consistency: The Full Workflow.

By Adam Morgan9 May 202610 min read
Seedance 2.0 Character Consistency: The Full Workflow

Getting consistent characters out of Seedance 2.0 takes more than a good prompt. Here is the full chaining method.

Why Character Consistency Breaks in Seedance 2.0

Article illustration

Seedance 2.0 made serious progress on identity drift. The DiT-based architecture and vendor-claimed improvements to facial, clothing, and style consistency mean multi-shot character work is genuinely more stable than it was in the 1.x era. But "more stable" is not "solved", and the failure modes that remain are predictable enough to plan around once you understand why they happen.

The content filter is the first culprit. Seedance 2.0 applies moderation at the prompt level, and it is sensitive to facial and body descriptors in ways that are not always intuitive. A straightforward phrase like "defined cheekbones, strong jawline" can trigger a partial block not because the content is problematic but because the filter interprets specific physical descriptors as potentially leading toward content it is trained to suppress. The result is not always a hard rejection. More often, the generation proceeds but the moderation layer quietly erodes the detail: accessories disappear, distinctive marks soften, or the face resolves into a generic average. You get a clip, but not your character.

Vague prompts compound this. When the model has insufficient information to resolve ambiguity, it defaults to generalisation, and the filter becomes more aggressive because generalised facial descriptors pattern-match more readily to flagged categories. Ironically, being less specific often makes the problem worse, not better.

The documented failure modes fall into three categories:

Drift Pattern What Happens Typical Trigger
Feature Erosion Accessories or distinctive marks vanish mid-clip Specific descriptor phrases flagged by filter
Pose Flip Hand dominance or gaze direction mirrors unexpectedly Ambiguous orientation in reference or prompt
Systemic Drift Face shape and proportions shift across shots No seed lock; prompt variation between scenes

These failures hit differently depending on what you are building. A game developer constructing a recurring NPC needs geometric stability across animation cycles: a face shape that drifts between idle and combat states breaks the illusion entirely. A fashion content creator generating a six-clip campaign series needs costume consistency that survives environment changes. A film team using Seedance for storyboard stand-ins needs the same actor readable across interior and exterior shots. The discipline differs. The wall is the same.

The critical reframe is this: consistency is a system problem, not a prompt problem. One well-crafted phrase may stabilise a single generation. Across a sequence of shots, without structural anchors, drift is inevitable. The rest of this guide is about building that structure.

Filter failures are rarely hard blocks. More often they are silent erosions: the clip renders, but your character arrives incomplete. Build your workflow to catch these early.

Build the Character Bible Before You Generate

Article illustration

The character bible is not a prompt. It is the source document that every prompt, reference image, and generation parameter derives from. Writing it before you open any generation surface forces the decisions that prevent drift downstream.

In Stensyl's Write studio, open a new document and structure it across four sections:

  1. Physical constants: Face structure described in environment and texture terms rather than direct physical descriptors. "Angular features, Mediterranean olive complexion, dark close-cropped hair" is more filter-stable than anatomical specifics.
  2. Clothing constants: Specific garments, palette hex codes where relevant, and fit descriptions. Note which costume details are non-negotiable and which can flex.
  3. Lighting and environment anchor: A reference lighting setup (e.g., "soft overcast daylight, no hard shadows, neutral background") that will serve as the canonical base for the character sheet.
  4. Banned ambiguities: A short list of phrases and constructions you have tested and found to trigger the filter or generate drift. This list grows as the project does.

Write is also where you stress-test the document before it reaches a generation surface. On the Pro tier, Claude Sonnet 4.6 and Claude Opus 4.7 are available in the model picker. Feed either model your drafted character description and ask it to identify which phrases are likely to conflict with video generation moderation layers, and to suggest filter-safe reformulations. Claude's large context window means you can paste an entire character bible and get back a line-by-line audit rather than a summary response.

The vocabulary differs by discipline, but the brief structure is identical. A product designer briefing a brand ambassador for a campaign video needs the same four sections as a game developer writing an NPC sheet. The product designer's "costume constants" section names the product colour family and brand-adjacent wardrobe cues. The game developer's version names armour weight class, faction insignia placement, and skin tone palette. Same document skeleton. The specificity of domain vocabulary is what changes.

Teams working in Stensyl Projects can store the character bible as a shared asset. When a client changes the costume mid-campaign (and they will), one update to the brief flows through to every team member's workflow rather than requiring six individual prompt edits across six assets.

A character bible stored in Projects is a living document, not a one-off prompt. Every hour spent sharpening it saves three hours of re-generation later.

The Moodboard-to-Generate Chain for Visual Anchoring

Prompts drift because language is imprecise. "Warm amber lighting" means something different to the model on a Tuesday afternoon than it did on Monday morning, and across a ten-shot sequence those small differences accumulate into visible inconsistency. A populated Moodboard in Stensyl functions as a visual constant that words cannot fully replace.

Open a Moodboard and populate it with three categories of reference:

  • Face structure proxies: Real photographs of faces with similar proportions to your character, ideally in the three-angle canonical arrangement (front, profile, three-quarter view) under neutral flat lighting. High-contrast features register more reliably as reference anchors.
  • Costume palette references: Fabric texture images, colour swatches, and full-length costume shots. Seedance's multimodal slots respond to image references for clothing, so specificity here pays dividends.
  • Environment and lighting tone: Location photography or lighting reference stills that establish the visual world the character inhabits. Even if your character appears in multiple environments, a canonical "home" lighting reference keeps the model's baseline consistent.

The Moodboard serves two practical functions. First, it is a return point between generation sessions. When you close the project and reopen it three days later, the Moodboard re-establishes the visual language immediately rather than requiring you to reconstruct it from memory or notes. Second, it creates a brief you can hand to a collaborator who needs to generate matching assets without a long verbal explanation.

When you move to Generate, use the Moodboard's established visual language as your prompt anchor. Reference the specific lighting conditions, the exact colour palette, and the costume textures you have already resolved visually. The goal is to reduce the model's interpretive latitude, not expand it.

This step is particularly high-value for motion designers building a character for an animated sequence across multiple scene contexts, where the character needs to read consistently whether they are in a studio, an exterior location, or a graphically stylised environment. Exhibition designers creating a recurring host figure across multiple installation screens face the same challenge: the character must be recognisable across wildly different display contexts and viewing distances. A well-populated Moodboard is the answer to both.

Sequencing Shots in Film and Storyboards Without Losing the Character

Article illustration

The Film surface in Stensyl is designed for multi-scene video work, and its structure is what prevents character descriptors from being re-invented from scratch between shots. The key principle is carryover: your character definition enters the sequence at the top and remains anchored throughout, rather than being reconstructed scene by scene.

Set up your scene sequence in Film with the character's core descriptors locked at the project level before you define individual scene prompts. Each scene then inherits those constants and adds only what is scene-specific: location, action, camera angle. This is a fundamentally different approach to prompting than treating each scene as an isolated generation task, and it dramatically reduces the variance that invites both drift and filter triggers.

Storyboards earns its place in this workflow before generation begins. Boarding the sequence first forces a specific commitment: you decide camera angle, framing, and shot type before the model has any input. This matters for consistency in a precise way. Many filter triggers and drift events are caused by prompts that are ambiguous about orientation. "Character walking forward" leaves gaze direction, dominant hand, and body angle unresolved. A boarded frame specifies all of these, and a prompt written against a boarded frame is structurally tighter and generates fewer moderation edge cases.

The practical chain is: board the full sequence in Storyboards first, then use the boarded frames as structural constraints when you move to Film. You are not describing what you hope the model will produce. You are describing what the board already shows.

One practical anchor point that reduces variability across the whole sequence: keep one filter-safe canonical shot as the first generation in every new scene group. A mid-shot, neutral expression, flat lighting, and no significant costume complexity. This "DNA check" generation confirms your descriptors are resolving correctly before you push into more complex camera angles or action states. If the canonical shot drifts, fix it at that level before building outward. Never build a sequence on a drifted base.

Board before you generate. A locked frame removes the ambiguity that causes filter triggers. Every prompt written against a specific board frame is a tighter, safer generation.

Using Canvas to Pipe Character Prompts Across the Workflow

Manual prompt copying is a consistency risk. Every time a character descriptor is retyped or partially recalled, the possibility of variation enters the workflow. Canvas, Stensyl's node-based editor, removes that risk by connecting a single character prompt source to multiple generation outputs.

The structure is straightforward. Place an LLM Chat node and connect it to a Video Generate node. In the LLM Chat node, you can select from the same writing models available in Write: Claude Sonnet 4.6 or Claude Opus 4.7 on Pro tier, or GPT-5.5 and Gemini Pro on Starter. Feed the node your character bible from Write (paste it in as context), then write an instruction asking it to reformat the core character descriptors into Seedance-safe prompt syntax: environment-based physical language, specific costume descriptions, and a lighting anchor. The output pipes directly into the Video Generate node.

What makes this worth building is the propagation behaviour. When the client changes the costume midway through production (and see above: they will), you update the character bible in the LLM Chat node once. Every connected Video Generate node receives the updated descriptor on the next run. You are not hunting through six separate prompt fields and trying to remember which ones you already updated.

For a marketing team producing a recurring spokesperson across six social video assets, this architecture maintains a genuine single source of truth for the character. The review cycle shortens because the costume change, the hair update, or the wardrobe swap propagates consistently rather than requiring a manual audit of every asset. The person checking the outputs is confirming consistency, not hunting for the places it broke.

Canvas also makes the workflow legible to collaborators. A node graph is a visible, inspectable representation of the character pipeline. A new team member can read it and understand where the character definition lives, how it reaches the generation nodes, and where to make changes. A folder of accumulated prompt strings offers none of that clarity.

Frame-Level Fixes in Editing When the Filter Still Wins

No workflow eliminates filter-triggered generations entirely. The filter will occasionally win. A specific phrase combination, a reference image that pattern-matches unexpectedly, or simply the stochastic nature of the model will produce a shot where the character has lost a defining feature, a clip where the face resolves as a generic stand-in, or a frame where an accessory simply is not there.

Stensyl's Editing surface (desktop only) is the right place to address these, with clear limits on what it can and cannot do. Editing works well for correcting partial renders: adjusting framing to reduce the prominence of a drifted feature, cleaning up compression artefacts or smearing at the edges of a face, and removing distracting frame-level inconsistencies that would otherwise force a full re-generation. It is not a reconstruction tool. If a key character feature is structurally absent from a frame, Editing cannot rebuild it from nothing. The decision point is: can the existing pixels be adjusted to an acceptable state, or does this shot need to be re-generated?

The triage approach that works in practice: flag failed frames during the Storyboards review pass, before you have invested time in a full edit. Isolate them. For each one, make a binary decision: fix in Editing, or regenerate with a tighter anchor. Only re-generate shots that cannot be salvaged. Re-generation is not failure; it is the right call when the fix time in Editing would exceed the time cost of running the shot again with corrected descriptors.

Each Stensyl surface has a defined job in this workflow:

  • Write: Drafts and stress-tests the character bible. The source of truth for all descriptors.
  • Moodboards: Establishes the visual anchor. Reduces prompt variance between sessions.
  • Storyboards: Locks shot composition before generation. Removes orientation ambiguity from prompts.
  • Film: Sequences scenes with inherited character constants. Prevents per-scene reinvention.
  • Canvas: Connects a single character source to multiple generation outputs. Propagates changes without manual updates.
  • Editing: Catches the frames that slip through. Corrects salvageable artefacts without re-running the full generation.

Character consistency across a multi-shot sequence is not the result of a single clever technique. It is the result of each surface doing its specific job, in sequence, with a character brief that was built to last rather than written on the fly. The failure modes are predictable. The workflow that counters them is buildable. Start with the bible.

Keep reading.

Try Stensyl for yourself

Image, video, 3D, chat, and document drafting. Every AI model, one studio. Plans from £10/month.