Community

AES Conference Papers Forum

A General Framework for Multichannel Speech Dereverberation Exploiting Sparsity

Document Thumbnail

We consider the problem of blind multi-channel speech dereverberation without the knowledge of room acoustics. The dereverberated speech component is estimated by subtracting the undesired component, estimated using multi-channel linear prediction (MCLP), from the reference microphone signal. In this paper we present a framework for MCLP-based speech dereverberation by exploiting sparsity in the time-frequency domain. The presented framework uses a wideband or a narrowband signal model and a sparse analysis or synthesis model for the desired speech component. The proposed problems involving a reweighted $\ell_1$-norm, are solved in a flexible optimization framework. The obtained results are comparable to the state of the art, motivating further extensions exploiting sparsity and speech structure.

Authors:
Affiliations:
AES Conference:
Paper Number:
Publication Date:
Subject:

Click to purchase paper as a non-member or you can login as an AES member to see more options.

No AES members have commented on this paper yet.

Subscribe to this discussion

RSS Feed To be notified of new comments on this paper you can subscribe to this RSS feed. Forum users should login to see additional options.

Start a discussion!

If you would like to start a discussion about this paper and are an AES member then you can login here:
Username:
Password:

If you are not yet an AES member and have something important to say about this paper then we urge you to join the AES today and make your voice heard. You can join online today by clicking here.

AES - Audio Engineering Society