Machine Learning

Conditional Generative Adversarial Networks for Speech Enhancement and Noise-Robust Speaker Verification

This topic contains 0 replies, has 1 voice, and was last updated by  arXiv 2 months, 2 weeks ago.


  • arXiv
    5 pts

    Conditional Generative Adversarial Networks for Speech Enhancement and Noise-Robust Speaker Verification

    Improving speech system performance in noisy environments remains a challenging task, and speech enhancement (SE) is one of the effective techniques to solve the problem. Motivated by the promising results of generative adversarial networks (GANs) in a variety of image processing tasks, we explore the potential of conditional GANs (cGANs) for SE, and in particular, we make use of the image processing framework proposed by Isola et al. [1] to learn a mapping from the spectrogram of noisy speech to an enhanced counterpart. The SE cGAN consists of two networks, trained in an adversarial manner: a generator that tries to enhance the input noisy spectrogram, and a discriminator that tries to distinguish between enhanced spectrograms provided by the generator and clean ones from the database using the noisy spectrogram as a condition. We evaluate the performance of the cGAN method in terms of perceptual evaluation of speech quality (PESQ), short-time objective intelligibility (STOI), and equal error rate (EER) of speaker verification (an example application). Experimental results show that the cGAN method overall outperforms the classical short-time spectral amplitude minimum mean square error (STSA-MMSE) SE algorithm, and is comparable to a deep neural network-based SE approach (DNN-SE).

    Conditional Generative Adversarial Networks for Speech Enhancement and Noise-Robust Speaker Verification
    by Daniel Michelsanti, Zheng-Hua Tan
    https://arxiv.org/pdf/1709.01703v2.pdf

You must be logged in to reply to this topic.