Real-Time Audio-Guided Multi-Face Reenactment

You are here

Top Reasons to Join SPS Today!

1. IEEE Signal Processing Magazine
2. Signal Processing Digital Library*
3. Inside Signal Processing Newsletter
4. SPS Resource Center
5. Career advancement & recognition
6. Discounts on conferences and publications
7. Professional networking
8. Communities for students, young professionals, and women
9. Volunteer opportunities
10. Coming soon! PDH/CEU credits
Click here to learn more.

Real-Time Audio-Guided Multi-Face Reenactment

By: 
Jiangning Zhang; Xianfang Zeng; Chao Xu; Yong Liu

Audio-guided face reenactment aims to generate authentic target faces that have matched facial expression of the input audio, and many learning-based methods have successfully achieved this. However, most methods can only reenact a particular person once trained or suffer from the low-quality generation of the target images. Also, nearly none of the current reenactment works consider the model size and running speed that are important for practical use. To solve the above challenges, we propose an efficient A udio-guided M ulti-face reenactment model named AMNet , which can reenact target faces among multiple persons with corresponding source faces and drive signals as inputs. Concretely, we design a Geometric Controller (GC) module to inject the drive signals so that the model can be optimized in an end-to-end manner and generate more authentic images. Also, we adopt a lightweight network for our face reenactor so that the model can run in real-time on both CPU and GPU devices. Abundant experiments prove our approach’s superiority over existing methods, e.g ., averagely decreasing FID by 0.12  and increasing SSIM by 0.031  than APB2Face, while owning fewer parameters ( ×4 ) and faster CPU speed ( ×4 ).

Audio-guided face reenactment aims to generate authentic target faces under the condition of audio information along with auxiliary pose and eye blink signals, which has promising applications such as animation production, virtual human, and game. However, most current methods can only reenact a particular person once finishing the training procedure or suffer from the low-quality problem of the generated target images. Also, nearly none of the current reenactment works take the model size and running speed into account that is important for practical use. This work focuses on solving the above problems, and we improve previous APB2Face [1] to an efficient end-to-end model to handle audio-guided multi-face reenactment, where different target faces among multiple persons can be reenacted by only one unified model.

SPS on Twitter

  • DEADLINE EXTENDED: The 2023 IEEE International Workshop on Machine Learning for Signal Processing is now accepting… https://t.co/NLH2u19a3y
  • ONE MONTH OUT! We are celebrating the inaugural SPS Day on 2 June, honoring the date the Society was established in… https://t.co/V6Z3wKGK1O
  • The new SPS Scholarship Program welcomes applications from students interested in pursuing signal processing educat… https://t.co/0aYPMDSWDj
  • CALL FOR PAPERS: The IEEE Journal of Selected Topics in Signal Processing is now seeking submissions for a Special… https://t.co/NPCGrSjQbh
  • Test your knowledge of signal processing history with our April trivia! Our 75th anniversary celebration continues:… https://t.co/4xal7voFER

IEEE SPS Educational Resources

IEEE SPS Resource Center

IEEE SPS YouTube Channel