首页|Adversarial multi-task learning with inverse mapping for speech enhancement

Adversarial multi-task learning with inverse mapping for speech enhancement

扫码查看
Adversarial Multi-Task Learning (AMTL) has demonstrated its promising capability of information capturing and representation learning, however, is hardly explored in speech enhancement. In this paper, we propose a novel adversarial multi-task learning with inverse mapping method for speech enhancement. Our method focuses on enhancing the generator's capability of speech information capturing and representation learning. To implement this method, two extra networks (namely P and Q) are developed to establish the inverse mapping from the generated distribution to the input data domains. Correspondingly, two new loss functions (i.e., latent loss and equilibrium loss) are proposed for the inverse mapping learning and the enhancement model training with the original adversarial loss. Our method obtains the state-of-the-art performance in terms of speech quality (PESQ=2.93, CVOL=3.55). For speech intelligibility, our method can also obtain competitive performance (STOI=0.947). The experimental results demonstrate that our method can effectively improve speech representation learning and speech enhancement performance. (c) 2022 Elsevier B.V. All rights reserved.

Speech enhancementAdversarial multi-task learningInverse mapping learningDeep neural networksINTELLIGIBILITY

Qiu, Yuanhang、Wang, Ruili、Hou, Feng、Singh, Satwinder、Ma, Zhizhong、Jia, Xiaoyun

展开 >

Massey Univ

2022

Applied Soft Computing

Applied Soft Computing

EISCI
ISSN:1568-4946
年,卷(期):2022.120
  • 4
  • 50