• Complex
  • Title
  • Keyword
  • Abstract
  • Scholars
  • Journal
  • ISSN
  • Conference
搜索

Author:

Xiang, Yang (Xiang, Yang.) | Bao, Changchun (Bao, Changchun.) | Yuan, Jing (Yuan, Jing.)

Indexed by:

EI Scopus

Abstract:

Nowadays, due to the application of deep neural network (DNNS), speech enhancement (SE) technology has been significantly developed. However, most of current approaches need the parallel corpus that consists of noisy signals, corresponding speech signals and noise on the DNNs training stage. This means that a large number of realistic noisy speech signals is difficult to train the DNNs. As a result, the performance of the DNNs is restricted. In this research, a new weakly supervised speech enhancement approach is proposed to break this restriction, using the cycle-consistent generative adversarial network (CycleGAN). There are two stage for our methods. In training stage, a forward generator is employed to estimate ideal time-frequency (T-F) mask and an inverse generator is utilized to acquire noisy speech magnitude spectrum (MS). Additionally, two discriminators are used to distinguish the real clean and noisy speech from generated speech, respectively. In enhancement stage, the T-F mask is directly estimated by using the well-trained forward generator for speech enhancement. Experimental results indicate that our strategy can not only achieve satisfied performance for non-parallel data, but also acquire the higher score in speech quality and intelligibility for the DNN-based speech enhancement using parallel data. © 2020 IEEE.

Keyword:

Deep neural networks Speech intelligibility Speech enhancement Generative adversarial networks Speech communication

Author Community:

  • [ 1 ] [Xiang, Yang]Beijing University of Technology, Beijing; 100124, China
  • [ 2 ] [Bao, Changchun]Beijing University of Technology, Beijing; 100124, China
  • [ 3 ] [Yuan, Jing]Beijing University of Technology, Beijing; 100124, China

Reprint Author's Address:

Email:

Show more details

Related Keywords:

Related Article:

Source :

Year: 2020

Language: English

Cited Count:

WoS CC Cited Count: 0

SCOPUS Cited Count: 2

ESI Highly Cited Papers on the List: 0 Unfold All

WanFang Cited Count:

Chinese Cited Count:

30 Days PV: 10

Affiliated Colleges:

Online/Total:477/10515955
Address:BJUT Library(100 Pingleyuan,Chaoyang District,Beijing 100124, China Post Code:100124) Contact Us:010-67392185
Copyright:BJUT Library Technical Support:Beijing Aegean Software Co., Ltd.