Character: Dr. Emmett Brown in "Back to the Future"
Vocal Source: Rick and Morty
we propose a novel audio-driven talking head method capable of simultaneously generating highly expressive facial expressions and hand gestures. Unlike existing methods that focus on generating full-body or half-body poses, we investigate the challenges of audio-driven ges- ture generation and identify the weak correspon- dence between audio features and full-body ges- tures as a key limitation. To address this, we re- define the task as a two-stage process. In the first stage, we generate hand poses directly from audio input, leveraging the stronger correlation between audio signals and hand movements. In the second stage, we employ a diffusion model to synthe- size video frames, incorporating the hand poses generated in the first stage to produce realistic facial expressions and body movements.
The motivation behind our method. Human motion, similar to that of robots, involves planning the "end-effector" (EE), typically the hands, towards the target situation. The rest of the body then cooperates accordingly with the EE, using inverse kinematics principles.
By inputting a single character image and vocal audio, such as singing, our method can generate vocal avatar videos featuring not only expressive facial expressions but also a variety of body poses.
Our method supports voice in multiple languages and brings images to life by intuitively recognizing tonal variations in audio, enabling the creation of dynamic, richly performing avatars.
Our method can generate complex and smooth hand movements, bringing the avatar to life with a vivid performance.
One potential application of our method is to enable designated characters to act out relevant scripts in film and game scenarios, with performances that align with their character profiles.
Check out our lighthearted video, created using our method. This video serves as a demonstration of potential application scenarios for our research. Hope you like it, and it will truely raise me up.