• 中国计算机学会会刊
  • 中国科技核心期刊
  • 中文核心期刊

Computer Engineering & Science ›› 2021, Vol. 43 ›› Issue (06): 1088-1094.

Previous Articles     Next Articles

Speech separation of cooperative training generative adversarial networks under low SNR

WANG Tao,QUAN Hai-yan   

  1. (Faculty of Information Engineering and Automation,Kunming University of Science and Technology,Kunming 650500,China)

  • Received:2019-12-02 Revised:2020-06-16 Accepted:2021-06-25 Online:2021-06-25 Published:2021-06-22

Abstract: Improving the quality of separated speech under low SNR is the focus of speech separation technology research, while most methods still only train the features of the target speaker's speech under low SNR. Aiming at the shortcoming of current methods, a mixed speech separation method based on cooperative training generative adversarial networks(GAN) is proposed. In order to avoid the extraction of the complex acoustic feature, the generative model uses a fully convolutional neural network to directly extract the high-dimensional features of the time-domain waveform, and the discriminative model obtains the features of the interference speaker by constructing a binary classification convolution neural network. Then, the source of the separated information obtained by the system is no longer single. Experiments show that the proposed method can better recover the information of high-frequency components under low SNR, and the separation performance is better than that of the comparative methods on the two-speaker mixed speech data set.



Key words: low SNR, generative adversarial networks, cooperative training, speech separation ,