Affiliation:
1. Department of Electrical Engineering, National Chi Nan University, Nantou 545, Taiwan
Abstract
Speech separation, sometimes known as the “cocktail party problem”, is the process of separating individual speech signals from an audio mixture that includes ambient noises and several speakers. The goal is to extract the target speech in this complicated sound scenario and either make it easier to understand or increase its quality so that it may be used in subsequent processing. Speech separation on overlapping audio data is important for many speech-processing tasks, including natural language processing, automatic speech recognition, and intelligent personal assistants. New speech separation algorithms are often built on a deep neural network (DNN) structure, which seeks to learn the complex relationship between the speech mixture and any specific speech source of interest. DNN-based speech separation algorithms outperform conventional statistics-based methods, although they typically need a lot of processing and/or a larger model size. This study presents a new end-to-end speech separation network called ESC-MASD-Net (effective speaker separation through convolutional multi-view attention and SuDoRM-RF network), which has relatively fewer model parameters compared with the state-of-the-art speech separation architectures. The network is partly inspired by the SuDoRM-RF++ network, which uses multiple time-resolution features with downsampling and resampling for effective speech separation. ESC-MASD-Net incorporates the multi-view attention and residual conformer modules into SuDoRM-RF++. Additionally, the U-Convolutional block in ESC-MASD-Net is refined with a conformer layer. Experiments conducted on the WHAM! dataset show that ESC-MASD-Net outperforms SuDoRM-RF++ significantly in the SI-SDRi metric. Furthermore, the use of the conformer layer has also improved the performance of ESC-MASD-Net.
Reference22 articles.
1. On training targets for supervised speech separation;Wang;IEEE/ACM Trans. Audio Speech Lang. Process.,2014
2. Hershey, J.R., Chen, Z., Roux, J.L., and Watanabe, S. (2016). Deep clustering: Discriminative embeddings for segmentation and separation. arXiv.
3. Isik, Y., Roux, J.L., Chen, Z., Watanabe, S., and Hershey, J.R. (2016). Single-channel multi-speaker separation using deep clustering. arXiv.
4. Luo, Y., and Mesgarani, N. (2018). Tasnet: Time-domain audio separation network for real-time, single-channel speech separation. arXiv.
5. Conv-TasNet: Surpassing ideal time–frequency magnitude masking for speech separation;Luo;IEEE/ACM Trans. Audio Speech Lang. Process.,2019
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献