Sound source localization and speech enhancement with sparse Bayesian learning beamforming

Angeliki Xenaki, Jesper Bünsow Boldt, Mads Græsbøll Christensen

Research output: Contribution to journalJournal articleResearchpeer-review

37 Citations (Scopus)
347 Downloads (Pure)


Speech localization and enhancement involves sound source mapping and reconstruction from noisy recordings of speech mixtures with microphone arrays. Conventional beamforming methods suffer from low resolution, especially with a limited number of microphones. In practice, there are only a few sources compared to the possible directions-of-arrival (DOA). Hence, DOA estimation is formulated as a sparse signal reconstruction problem and solved with sparse Bayesian learning (SBL). SBL uses a hierarchical two-level Bayesian inference to reconstruct sparse estimates from a small set of observations. The first level derives the posterior probability of the complex source amplitudes from the data likelihood and the prior. The second level tunes the prior towards sparse solutions with hyperparameters which maximize the evidence, i.e., the data probability. The adaptive learning of the hyperparameters from the data auto-regularizes the inference problem towards sparse robust estimates. Simulations and experimental data demonstrate that SBL beamforming provides high-resolution DOA maps outperforming traditional methods especially for correlated or non-stationary signals. Specifically for speech signals, the high-resolution SBL reconstruction offers not only speech enhancement but effectively speech separation.
Original languageEnglish
JournalThe Journal of the Acoustical Society of America
Issue number6
Pages (from-to)3912-3921
Number of pages10
Publication statusPublished - Jun 2018


Dive into the research topics of 'Sound source localization and speech enhancement with sparse Bayesian learning beamforming'. Together they form a unique fingerprint.

Cite this