We consider the problem of blind multi-channel speech dereverberation without the knowledge of room acoustics. The dereverberated speech component is estimated by subtracting the undesired component, estimated using multi-channel linear prediction (MCLP), from the reference microphone signal. In this paper we present a framework for MCLP-based speech dereverberation by exploiting sparsity in the time-frequency domain. The presented framework uses a wideband or a narrowband signal model and a sparse analysis or synthesis model for the desired speech component. The proposed problems involving a reweighted $\ell_1$-norm, are solved in a flexible optimization framework. The obtained results are comparable to the state of the art, motivating further extensions exploiting sparsity and speech structure.
Authors:
Jukic, Ante; van Waterschoot, Toon; Gerkmann, Timo; Doclo, Simon
Affiliations:
KU Leuven, Leuven, Belgium; University of Oldenburg, Oldenburg, Germany(See document for exact affiliation information.)
AES Conference:
60th International Conference: DREAMS (Dereverberation and Reverberation of Audio, Music, and Speech) (January 2016)
Paper Number:
9-1
Publication Date:
January 27, 2016
Subject:
Paper Session 9
Click to purchase paper as a non-member or you can login as an AES member to see more options.
No AES members have commented on this paper yet.
To be notified of new comments on this paper you can
subscribe to this RSS feed.
Forum users should login to see additional options.
If you are not yet an AES member and have something important to say about this paper then we urge you to join the AES today and make your voice heard. You can join online today by clicking here.