Emote Portrait Alive: Generating Expressive Portrait Videos with Audio2Video Diffusion Model under Weak Conditions

Emote Portrait Alive: Generating Expressive Portrait Videos with Audio2Video Diffusion Model under Weak Conditions

EMO is an audio-driven portrait-to-video generation framework. Input a single reference image and the vocal audio, e.g. talking and singing, and generate vocal avatar videos with expressive facial expressions, various head poses. Generate any duration depending on the length of input video.

You can make still images talk, sing, or cross-actor operations in which still images can deliver the performances of other actors or in different languages.

It reminds me of other AI facial animation software like MyHeritage’s Deep Nostalgia.

Project: https://humanaigc.github.io/emote-portrait-alive/

Github: https://github.com/HumanAIGC/EMO

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.