Speech Enhancement based on Denoising Autoencoder with Multi-branched Encoders

01/06/2020
by   Cheng Yu, et al.
0

Deep learning-based models have greatly advanced the performance of speech enhancement (SE) systems. However, two problems remain unsolved, which are closely related to model generalizability to noisy conditions: (1) mismatched noisy condition during testing, i.e., the performance is generally sub-optimal when models are tested with unseen noise types that are not involved in the training data; (2) local focus on specific noisy conditions, i.e., models trained using multiple types of noises cannot optimally remove a specific noise type even though the noise type has been involved in the training data. These problems are common in real applications. In this paper, we propose a novel denoising autoencoder with a multi-branched encoder (termed DAEME) model to deal with these two problems. In the DAEME model, two stages are involved: offline and online. In the offline stage, we build multiple component models to form a multi-branched encoder based on a dynamically-sized decision tree(DSDT). The DSDT is built based on a prior knowledge of speech and noisy conditions (the speaker, environment, and signal factors are considered in this paper), where each component of the multi-branched encoder performs a particular mapping from noisy to clean speech along the branch in the DSDT. Finally, a decoder is trained on top of the multi-branched encoder. In the online stage, noisy speech is first processed by the tree and fed to each component model. The multiple outputs from these models are then integrated into the decoder to determine the final enhanced speech. Experimental results show that DAEME is superior to several baseline models in terms of objective evaluation metrics and the quality of subjective human listening tests.

READ FULL TEXT

page 1

page 6

page 7

page 9

research
03/22/2022

Joint Noise Reduction and Listening Enhancement for Full-End Speech Enhancement

Speech enhancement (SE) methods mainly focus on recovering clean speech ...
research
07/19/2018

Noise Adaptive Speech Enhancement using Domain Adversarial Training

In this study, we propose a novel noise adaptive speech enhancement (SE)...
research
02/17/2021

Variational Autoencoder for Speech Enhancement with a Noise-Aware Encoder

Recently, a generative variational autoencoder (VAE) has been proposed f...
research
09/17/2023

Continuous Modeling of the Denoising Process for Speech Enhancement Based on Deep Learning

In this paper, we explore a continuous modeling approach for deep-learni...
research
10/12/2021

MetricGAN-U: Unsupervised speech enhancement/ dereverberation based only on noisy/ reverberated speech

Most of the deep learning-based speech enhancement models are learned in...
research
11/07/2018

On the use of DNN Autoencoder for Robust Speaker Recognition

In this paper, we present an analysis of a DNN-based autoencoder for spe...
research
01/12/2018

Speech Dereverberation Based on Integrated Deep and Ensemble Learning

Reverberation, which is generally caused by sound reflections from walls...

Please sign up or login with your details

Forgot password? Click here to reset