Experimental music video merging live-action with AI-animated characters while preserving performance fidelity
Following the innovative path laid by my Ex Machina AI rendering experiment, I undertook a new project to explore the evolving landscape of technology in the creative field. "Immersion: Real Meets AI" is an experimental music video that aims to increase performance fidelity in AI filmmaking by merging live-action footage with 2D animation, inspired by the style of "Who Framed Roger Rabbit."
The twist in this project was integrating AI to animate the 2D character, creating a unique blend of human and artificial creativity. This endeavor allowed me to develop and refine a new video-to-video (vid2vid) workflow, pushing the limits of how AI can stylize and render existing videos into animated sequences. The project showcased my compositing and VFX skills while allowing me to practice model fine-tuning to preserve the actress's likeness and expressivity.
What makes this project truly innovative is the character consistency I achieved by combining a trained LoRA on my actress with IP adapter technology, solving one of the most persistent challenges in AI-generated video. Through "Immersion: Real Meets AI," I aimed to bridge the gap between traditional filmmaking and cutting-edge AI technology, opening new avenues for creativity and storytelling in the digital age.
Create a new AI workflow that preserves character consistency and performance fidelity while merging live-action and animated elements.
4 months (August 2023 - December 2023), developed during USC's "Directing in a Virtual World" class.
Creator, Director, VFX Artist, AI Engineer, Compositor, and Technical Director for the entire workflow.
ComfyUI, AnimateDiff, Green Screen, Stable Diffusion, LoRA Training, IP Adapter, After Effects, HDRIs, Video-to-Video Processing.
Creating a believable interaction between live-action and AI-generated animated characters posed several significant technical and creative challenges:
I developed a comprehensive approach that addressed each challenge with innovative technical and creative solutions:
This project evolved significantly throughout its development, as I adapted to rapidly changing AI technologies and refined my approach based on results. What began as an evolution of my Ex Machina workflow transformed into something entirely new as I incorporated cutting-edge tools like AnimateDiff and developed innovative filming techniques.
The project began as an experiment for USC's "Directing in a Virtual World" class. My initial idea evolved from using the LED wall as a rough background with AI rendering for polish, to using it as a physical barrier between characters, and finally to a "Who Framed Roger Rabbit" inspired approach with green screen. Working with my actress, we developed a story about a demon from another dimension who crash-lands on Earth, allowing for organic interaction between the characters.
Left: Early storyboard version; Right: Initial concept image showing the visual style
I approached character creation by first exploring using a 3D metahuman of my actress that could be processed with AI. I also developed a novel ComfyUI workflow that incorporated nodes for character consistency and expressivity preservation. When AnimateDiff was released during production, I quickly adapted my workflow to incorporate this powerful new technology, significantly improving the quality of the animated character.
Before/After: Metahuman of the actress before and after AI processing
Training data: Creative pose
Training data: Front facing reference
Early LoRA result
Final LoRA result with improved quality
LoRA training progression: From varied training data to increasingly refined results, showing how character consistency was achieved
Technical walkthrough of the ComfyUI workflow
I implemented a specialized green screen setup with laser dot tracking markers that provided precise tracking without affecting the green screen key. To capture realistic interactions, I used specific marks, C-stands, and a transparent piece of acrylic where actors could place their hands. I filmed the actors both together and separately for maximum flexibility, allowing for genuine interaction while still having clean isolated takes that could be processed individually.
Crew working on the green screen set
For the crucial eye reflection scene, instead of using post-production compositing, I projected a pre-rendered idle "animation" of the demon girl onto a large screen and filmed the actor looking at that projection, creating a genuine reflection in his eye. I also trained a LoRA model on my actress's face and combined it with IP adapter technology to achieve unprecedented character consistency. The backgrounds were created using HDRIs with subtle AI-generated movement to enhance the sense of immersion.
Before/after test showing transformation
Left: Actress with tracking points; Right: Processed result showing character consistency
The final result was a revolutionary approach to character animation that bridges traditional VFX techniques with cutting-edge AI technology. "Immersion: Real Meets AI" successfully demonstrated a new workflow for creating animated characters with consistent appearance and preserved performance fidelity, while substantially reducing the time and resources traditionally required.
Beyond the technical achievements, this project delivered several significant creative and professional results:
Actress closeup used for IP adapter reference
3D scan of the actress used for metahuman in early tests, and as training data
The technological breakthrough in this project came from the unique combination of several cutting-edge AI techniques and traditional filmmaking methods. Below is a detailed exploration of the key technical innovations that made this project possible.
Unlike my previous Ex Machina project which used Auto1111, this project utilized ComfyUI's node-based approach to create a more flexible and powerful workflow. The workflow incorporated:
Early test of character interaction showing the progression of the technique
The production utilized several innovative approaches to filming that enhanced the final result:
Left: Testing green screen with stand-in; Right: Additional green screen test from the same production
The final stage of the process involved several technical innovations in compositing:
This project represented a significant evolution in my exploration of AI for creative filmmaking. The rapid pace of AI development during the project period—including the release of AnimateDiff—created both challenges and opportunities that shaped the final outcome in ways I couldn't have anticipated at the outset.
This project reinforced my belief that the most powerful applications of AI in creative fields come not from replacing traditional techniques, but from creating new workflows that combine the best of both worlds. By using AI to handle the labor-intensive aspects of animation while preserving the human performance at its core, we can create expressive characters with unprecedented efficiency. The positive reception from industry professionals, particularly regarding the character consistency techniques, validates this approach and suggests exciting possibilities for the future of filmmaking.