Under-determined blind speech separation via the convolutive transfer function and lp Regularization

Liu Yang, Junjie Yang, Yi Guo

Research output: Chapter in Book / Conference PaperConference Paperpeer-review

1 Citation (Scopus)

Abstract

![CDATA[Convolutive blind speech separation (CBSS) attempts to recover speech sources from recorded mixtures transmitted by room impulse response filters (RIRs). The existing BSS methods are mainly based on a multiplicative narrowband approximation (MNA) model in the short-time Fourier transform (STFT) domain. However, the MNA model may result in serious system transformation error under the situation of a strongly reverberant environment due to its sensitivity to RIRs with large lengths. Moreover, the solution of source reconstruction is not unique when the system is underdetermined even if the RIRs are given in advance. To tackle these problems, a convolutive narrowband approximation (CNA) model is explored in the proposed approach to constrain the system transformation error. Then an l p(0lt pleq 1) regularization is provided to reconstruct the sparse sources in which p can be dynamically selected to achieve the best performance. The experimental results demonstrate the robustness of the proposed method to room reverberation under various speech separation cases in comparison to conventional methods.]]
Original languageEnglish
Title of host publicationProceedings of 17th International Conference on Mobility, Sensing and Networking, MSN 2021, 13-15 December 2021, Exeter, United Kingdom
PublisherIEEE
Pages705-709
Number of pages5
ISBN (Print)9781665406680
DOIs
Publication statusPublished - 2021
EventInternational Conference on Mobility_Sensing and Networking -
Duration: 13 Dec 2021 → …

Conference

ConferenceInternational Conference on Mobility_Sensing and Networking
Period13/12/21 → …

Fingerprint

Dive into the research topics of 'Under-determined blind speech separation via the convolutive transfer function and lp Regularization'. Together they form a unique fingerprint.

Cite this