Enhancing Facial Expressivity in AI Rendering

Pioneering a technique to preserve emotional performance in AI-enhanced 3D animation

Comparison showing DVA from Overwatch cinematic before and after AI rendering enhancement for facial expressivity

Project Overview

As part of my comprehensive investigation into AI-enhanced visual media at USC's School of Cinematic Arts, I conducted a critical experiment to address a fundamental challenge: preserving facial expressivity while improving visual quality. After observing how standard AI enhancement flattened emotions in facial performances, I developed a specialized workflow using a highly expressive Overwatch character (DVA) as my test subject, isolating facial processing from body rendering.

Working in early 2023, when generative AI was still in its infancy and dedicated video models didn't exist, I crafted innovative solutions that pushed the boundaries of what was then possible with image-based AI models. This experiment established key techniques that directly informed my approach to recreating scenes from Ex Machina and later character-focused projects, establishing a foundation for preserving performance quality in AI-enhanced rendering.

🎯

Goal

Enhance the realism of 3D animation while preserving the emotional fidelity of facial performances

⏱️

Timeline

April 2023, during the early days of AI-enhanced rendering techniques

🧠

Role

Lead Researcher & Developer, designing and implementing the entire workflow

🛠️

Tools & Technologies

Stable Diffusion 1.5, Auto1111 WebUI, ControlNet, MediaPipe Face Detection, After Effects, DaVinci Resolve, Topaz Video AI

Challenge & Solution

The Challenge

During my exploration of AI rendering techniques, I encountered a significant obstacle: while AI enhancement provided impressive visual improvement to 3D animations, it came at the cost of emotional fidelity. Facial expressions became rigid and emotionless, defeating the purpose of performance capture and reducing the narrative impact.

  • AI models tended to "average out" subtle facial movements, resulting in flattened expressions
  • The parameters that worked well for body rendering degraded facial detail
  • Processing the entire frame with uniform settings created inconsistent results
  • Early 2023 AI tools lacked dedicated solutions for facial expression preservation

The Solution

I developed a novel dual-processing approach where I would process faces and bodies separately with different parameter sets, then composite them back together. This allowed me to optimize the parameters for facial expressivity without compromising the quality of the rest of the render.

  • Created a specialized face mask generation process using MediaPipe Face Detection
  • Processed faces with reduced denoising strength (0.5 vs. 0.65) to preserve movement details
  • Implemented dedicated ControlNet processing specifically for facial features
  • Developed a compositing workflow to seamlessly integrate the separately processed elements

Before & After: DVA Facial Expressivity Test - Note the preserved emotion and facial detail

Process & Methodology

I approached this challenge methodically, breaking down the problem into discrete steps that could be solved individually before being integrated into a cohesive workflow.

1

Problem Identification & Analysis

I conducted systematic testing to understand exactly how facial expressivity was being lost in AI rendering. Through careful parameter testing, I discovered that the denoising strength needed for photorealism was too aggressive for preserving subtle facial movements, and that the AI prioritized overall frame coherence over local detail preservation.

2

Face Mask Generation Workflow

I developed a specialized process for generating accurate face masks from each frame. Using MediaPipe Face Detection with specific parameters, I created masks that precisely isolated facial areas for targeted processing.

Face Detection Settings:
- Detector: Normal (OpenCV + FaceMesh)
- Minimum face size in pixels: 30
- Rotation threshold: 20
- Single mask per image: Enabled
3

Face-Specific Parameter Optimization

I conducted extensive experimentation to find the optimal parameters for facial processing. The key discovery was that reducing the denoising strength to 0.5 (from 0.65 used for bodies) preserved significantly more expressivity while still enhancing visual quality.

Inpaint Settings for Faces:
- Mask mode: Inpaint masked
- Masked content: Original
- Inpaint area: Only masked
- Only masked padding: 32px
- Sampling method: Euler
- Sampling steps: 30
- CFG Scale: 8
- Denoising strength: 0.5

ControlNet Settings:
- Preprocessor: mediapipe_face
- Model: control_v1p_sd15_mediapipe_face
- Weight: 1
- Guidance: Full range (0-1)
4

Integration & Compositing

The final stage involved tracking and compositing the enhanced faces onto the body renders. This required precise alignment and seamless blending to ensure natural integration. I further refined the process with frame interpolation techniques to improve smoothness and reduce flicker.

As I noted in my journal (April 17, 2023): "I need to try it and include it in the workflow... imagine: my current method plus three controlnets, one for body, one for depth/bg, and one for faces. All processing for each batched image."

Results & Impact

This experiment yielded remarkable results that not only solved the immediate problem but established foundational techniques that would inform my future AI rendering work.

100%
Expression Preservation
Successfully maintained all key facial expressions while enhancing visual quality
1st
SCA AI Project
First-ever USC School of Cinematic Arts project to utilize AI rendering techniques
0.15
Denoising Difference
Critical parameter difference between body and face processing that enabled success

Qualitative Outcomes

Beyond the technical metrics, this experiment achieved several significant outcomes:

  • Workflow Foundation: Established crucial rendering settings and techniques that would inform future projects
  • Technical Innovation: Developed a novel approach to AI rendering that preserved performance while enhancing visual quality
  • Professional Recognition: The technique impressed VFX veterans with its realistic subsurface scattering in hair and nuanced facial detail preservation
  • Historical Significance: Pioneered AI techniques in film production at USC, opening doors for future exploration

"The preservation of subtle facial movements while still achieving photorealistic enhancement demonstrates a sophisticated understanding of both AI models and character animation requirements."

— Faculty assessment of the DVA facial expression technique

Reflection & Learnings

This project represented a critical juncture in my exploration of AI rendering techniques, yielding insights that would shape my approach to subsequent work.

What Worked Well

  • Task Specialization: Focusing the AI on single, specialized tasks led to dramatically better results than trying to process everything with one set of parameters
  • Parameter Sensitivity: Small adjustments to key parameters (like denoising strength) had outsized impacts on quality
  • Iterative Testing: Methodical testing and documenting of results allowed me to isolate effective techniques

Challenges & Solutions

  • Detail Preservation: Achieved a delicate balance between enhancing detail and preserving natural movement
  • Compositing Alignment: Developed careful tracking techniques to ensure faces remained aligned with bodies
  • Processing Time: Optimized workflows to handle the significant processing time required for dual-processing

Future Applications

  • Extended Applications: Applied similar masked processing techniques to hands, which had similar issues with AI distortion
  • Interpolation Trade-offs: Discovered that while frame interpolation could improve smoothness, it often reduced expressivity, leading to a design choice to preserve emotion over technical perfection
  • Technical Evolution: Realized this technique would inform my subsequent character rendering projects

Personal Takeaway

This project taught me the importance of problem decomposition in AI—breaking complex challenges into specialized tasks. Working at the cutting edge of AI in early 2023, with minimal documentation and precedent, forced me to develop innovative approaches and trust my experimental results. The recognition from professors and industry professionals validated my unconventional approach and encouraged me to continue pushing boundaries in subsequent projects.