How to Replace a Character in a Video with AI
From plate preparation to believable swaps: what AI replacement can do today, how to shoot or source footage, and quality checks before delivery.
8 min read
Replacing a character in a video once required meticulous VFX: match-moving, roto, lighting reprojection, and compositing. AI-assisted replacement accelerates parts of that pipeline by synthesizing new appearance conditioned on motion that already exists in the plate. It is not magic—lighting, occlusion, and physics still matter—but when used with disciplined footage, results can be convincing and far faster than frame-by-frame paint.
Understand what the model is actually swapping
Most character replacement systems infer identity from a reference image and motion from the underlying performance or camera move. They are not always interpreting full 3D scene geometry. That means shots with clear silhouettes, stable framing, and minimal motion blur tend to succeed. Heavy occlusion—hands over face, props crossing the body—may require simpler alternate takes or traditional cleanup.
Shoot or select source video with replacement in mind
Favor moderate focal lengths; extreme lenses distort faces in ways that confuse synthesis. Keep exposure consistent within a clip; sudden exposure ramps fight relighting models. If possible, capture a clean plate or consistent background to aid edge reconstruction. When sourcing stock, avoid clips with aggressive color grading unless you will re-grade the final composite yourself.
Prepare your reference image
Use a high-quality still of the replacement character under lighting similar to the scene’s key light direction—even approximate alignment reduces shadow contradictions. Match wardrobe intent; if the plate shows a collared shirt, a reference in a heavy coat may produce fabric hallucinations. For stylized characters, expect to stylize the plate consistently; mixing photoreal video with a flat cartoon still without adjustment rarely holds.
Camera motion versus performer motion
Handheld micro-jitter can be charming in-camera but challenging for synthesis because the subject and background shift together in complex ways. Stabilizing plates mildly—without destroying parallax—sometimes improves results. Conversely, locked-off shots can expose temporal inconsistencies because viewers study the face without distraction. Choose your battles: stabilize when the tool struggles with shake; add subtle camera drift in post if the shot feels too clinical after replacement.
Multi-character scenes
When two actors overlap, masks and depth ordering become critical. AI may confuse limbs at contact points. Consider simplifying blocking, using wider shots, or splitting composites so each character processes with clearer silhouettes. Traditional roto may still be required at overlaps; plan budget for hybrid shots instead of assuming full automation.
Run tests on the hardest frames
Before processing an entire timeline, identify the worst moments: fast turns, profile-to-three-quarter transitions, and overlapping actors. Test those segments first. If the system fails on peaks, it will not magically improve on easier frames once assembled. Iterate on reference choice and prompt constraints until peaks are acceptable, then batch the full clip.
Edit for continuity
Even strong AI outputs benefit from editorial rhythm. Cut around lingering artifact frames, add reaction shots that do not stress the model, and use audio continuity to sell performance when visuals are slightly soft. Color match foreground and background so the replacement does not feel sticker-like. Small defocus or grain matched across layers integrates seams.
Ethics, consent, and disclosure
Replacing identifiable people without permission crosses legal and ethical lines in many jurisdictions. Use synthetic actors, licensed talent, or consented references. For public communications, disclose synthetic or altered media when regulations or platforms require it. Build review steps for brand safety, especially in news-adjacent or political contexts.
Using Wan Animate for replacement workflows
Wan Animate includes Replace-oriented flows alongside Animate, aimed at teams that need to substitute a character while preserving motion structure from the original clip. Treat it as part of a hybrid pipeline: AI for heavy lifting, editorial and compositing for control. The winning shots combine good plates, thoughtful references, and human judgment about when to trust the model—and when to cut away.
Deliverables checklist
Before handoff, verify frame rate and resolution match the master, check audio sync after retiming, and archive prompts plus source references for reproducibility. Run QC on target displays, including mobile. Document known limitations for stakeholders so expectations stay aligned. With disciplined inputs, AI character replacement becomes a practical production tool rather than a demo trick.
Audio and performance
If dialogue or song drives the scene, align mouth shapes loosely rather than expecting perfect phoneme accuracy from every system. ADR, foley, and music beds help sell a swap when the visuals are slightly imperfect. Editors who cut on rhythm often report happier stakeholders than those who chase frame-perfect lipsync on aggressive dialogue without specialized audio-visual models.
Archiving for reshoots and sequels
Store the original plate, the replacement reference, export settings, and any mask assets together in versioned folders. Sequels and pickups happen months later; future you will not remember which prompt saved the cheek highlight. A modest archive discipline turns a one-off AI experiment into a reusable pipeline the next time a director asks for “the same trick, but faster,” with fewer surprises for post supervisors.