Robust Multichannel Linear Prediction for Online Speech Dereverberation Using Weighted Householder Least Squares Lattice Adaptive Filter

You are here

Top Reasons to Join SPS Today!

1. IEEE Signal Processing Magazine
2. Signal Processing Digital Library*
3. Inside Signal Processing Newsletter
4. SPS Resource Center
5. Career advancement & recognition
6. Discounts on conferences and publications
7. Professional networking
8. Communities for students, young professionals, and women
9. Volunteer opportunities
10. Coming soon! PDH/CEU credits
Click here to learn more.

Robust Multichannel Linear Prediction for Online Speech Dereverberation Using Weighted Householder Least Squares Lattice Adaptive Filter

By: 
Jason Wung; Ante Jukić; Sarmad Malik; Mehrez Souden; Ramin Pichevar; Joshua Atkins; Devang Naik; Alex Acero

Speech dereverberation has been an important component of effective far-field voice interfaces in many applications. Algorithms based on multichannel linear prediction (MCLP) have been shown to be especially effective for blind speech dereverberation and numerous variants have been introduced in the literature. Most of these approaches can be derived from a common framework, where the MCLP problem for speech dereverberation is formulated as a weighted least squares problem that can be solved analytically. Since conventional batch MCLP-based dereverberation algorithms are not suitable for low-latency applications, a number of online variants based on the recursive least squares (RLS) algorithm have been proposed. However, RLS-based approaches often suffer from numerical instability and their use in online systems can further be limited due to high computational complexity with a large number of channels or filter taps. In this paper, we aim to address the issues of numerical robustness and computational complexity. More specifically, we derive alternative online weighted least squares algorithms through Householder RLS and Householder least squares lattice (HLSL), which are numerically stable and retain the fast convergence capability of the RLS algorithm. Furthermore, we derive an angle-normalized variant of the HLSL algorithm and show that it is robust to speech cancellation for a wide range of forgetting factors and filter taps. Finally, we support our findings through experimental results and demonstrate numerical and algorithmic robustness, long-term stability, linear complexity in filter taps, low memory footprint, and effectiveness in speech recognition applications.

SPS on Twitter

  • DEADLINE EXTENDED: The 2023 IEEE International Workshop on Machine Learning for Signal Processing is now accepting… https://t.co/NLH2u19a3y
  • ONE MONTH OUT! We are celebrating the inaugural SPS Day on 2 June, honoring the date the Society was established in… https://t.co/V6Z3wKGK1O
  • The new SPS Scholarship Program welcomes applications from students interested in pursuing signal processing educat… https://t.co/0aYPMDSWDj
  • CALL FOR PAPERS: The IEEE Journal of Selected Topics in Signal Processing is now seeking submissions for a Special… https://t.co/NPCGrSjQbh
  • Test your knowledge of signal processing history with our April trivia! Our 75th anniversary celebration continues:… https://t.co/4xal7voFER

IEEE SPS Educational Resources

IEEE SPS Resource Center

IEEE SPS YouTube Channel