融合面部深度感知的音频驱动人脸重现方法
Audio Driven Face Reenactment Method Integrating Facial Deep-perception
彭雪康 1孙国庆 1邵长乐 1练智超1
作者信息
- 1. 南京理工大学网络空间安全学院,南京 210094
- 折叠
摘要
人脸重现是一项条件面部生成任务,现有的基于音频驱动的人脸重现方法难以生成完整且高质量的人脸.针对这一问题,提出一种融合面部深度信息的音频驱动下的人脸重现方法.该方法采用了轻量级的模型框架以降低模型尺寸和提高运行速度.实验在AnnVI数据集上与3种最新的音频驱动人脸重现方法进行了比较.结果表明,所提出的融合面部深度感知的人脸重现方法,极大地提高了音频驱动下生成人脸图像的质量.
Abstract
Face reenactment is a conditional facial generation mission.The current face reenactment method based on audio driven is difficult to generate a complete and high quality face.As for such a problem,the audio driven face reenactment method integrating facial deep information is proposed.Meanwhile,the method adopts a light-weight model framework to reduce the model size and to improve the running speed.The experiment is compared with three kinds of the latest audio driving face reenactment methods on the AnnVI dataset.The results show that the proposed face reenactment method,integrating facial deep perception can greatly improve the quality of generated facial images driven by the audio.
关键词
人脸伪造/人脸重现/深度估计/多模态驱动/生成对抗网络Key words
deepfake/face reenactment/deep estimation/multi-modal driving/generative adversarial network引用本文复制引用
基金项目
国家重点研发计划(2021YFF0602104-2)
出版年
2024