Live Performance | Live Motion Capture | Projection Mapping | Virtual Production | Volumetric Capture
A hybrid performance and film where audience-driven physical control transforms a performer’s digital avatar in real-time, exploring the tension between identity, agency, and external influence.
The Morph is a hybrid performance and film that explores how external forces shape identity, agency, and embodiment.
In the live performance, a central performer stands in the middle of the space, physically restrained by ropes held by the audience. As they pull and drag the performer, her digital avatar—mirroring her movements—triggers real-time scene, lighting, and avatar transformations in a game engine. The avatar, initially featureless, gradually accumulates growing elements—representing external influences such as technology, power, societal expectations, or abstract organic mutations.
As the performance progresses, the performer’s physical and emotional state shifts; at first, she simply reacts, then begins to embrace her form, but as more is imposed upon her, the body becomes overburdened and ultimately collapses. In the final stage, the performer resists being pulled, struggling against the force of the audience yet unable to escape.
The work merges physical and digital space, making the performer both the subject of control and the emotional core of the avatar’s transformation.
The accompanying film, created with virtual production LED walls and volumetric capture, captures the most refined choreography from the performance, presenting a cinematic reimagination of the evolving body and its struggle.
To request access to the full film, please contact the artists for a private viewing link.
CREDITS:
Live Performance:
Creative Director:
John Luo Choregraphers:
Audrey Chou
Mingjun Han
User Experience Co Director:
Lois He
Projection Mapping:Damon Chen 3D Generalist:Damon Chen
John Luo
Developer: John LuoPerformer: Audrey Chou Music Composer: Audrey Chou
Special Thanks: NYU @Media Commons
Kevin Cunningham
Film:
Director
Performer
Choreographer
Sound Design:
Audrey Chou Technical Director
Unreal Engine
Virtual Production
Volumetric Capture:
Adi Somaiah
Writer
Character Design
Motion Capture
Unreal Engine:John Luo
Director of Photography
Unreal Engine
Video Editing
Gaffer: Matt Toups3D Generalist: John Luo
Damon Chen
Colorist Sania Bhatia
Special Thanks: NYU Tandon @ The Yard
WLAB XR
Matthew Rader
Evercoast
THE MORPH: LIVE PERFORMANCE
THE MORPH: FILM
LIVE PERFORMANCE: RIGIDBODY MOTION CAPTURE
The performance features a custom motion capture system developed in Unreal Engine, utilizing 3D-printed rigid body holders to track the dancer’s limbs as separate units. This setup enables precise motion capture that fragments the performer’s body, creating a unique visual effect where each limb moves independently yet remains connected. The system enhances the surreal transformation of the avatar, reinforcing themes of external control and evolving embodiment.
LIVE PERFORMANCE: PROJECTION MAPPING & LIVE INTERACTION
The performance took place in two distinct venues, each experimenting with audience perspective and interactivity. One featured projection mapping on a single long horizontal wall, guiding a focused line of sight, while the other surrounded the space with three walls projecting the same content. This setup allowed for an exploration of how spatial immersion and audience movement influenced engagement as they physically manipulated the performer.
FILM: MOTION CAPTURE & VOLUMETRIC CAPTURE
The film integrates both OptiTrack motion capture and volumetric capture to bring the creature to life while preserving the precision of the performer’s choreography. OptiTrack ensures accurate body movement tracking, while volumetric capture seamlessly embeds the dancer into the digital world, blending physical performance with the evolving avatar in a cinematic reimagination of transformation and control.
FILM: VIRTUAL PRODUCTION
The film utilizes virtual production with an LED wall, ensuring the digital world seamlessly integrates with practical lighting for realistic in-camera visuals. This setup enhances immersion for the performer, allowing them to react naturally to their environment while achieving accurate reflections, interactive lighting adjustments, and real-time scene changes. It also reduces post-production compositing, creating a more cohesive and efficient workflow.