Reverberation is ubiquitous in everyday listening environments, from meeting rooms to concert halls and record-ing studios. While reverberation is usually described by the reverberation time, getting further insight concerning the characteristics of a room requires to conduct acoustic measurements and calculate each reverberation param-eter manually. In this study, we propose ReverbNet, an end-to-end deep learning-based system to non-intrusively estimate multiple reverberation parameters from a single speech utterance. The proposed approach is evaluated using simulated room reverberation by two popular effect processors. We show that the proposed approach can jointly estimate multiple reverberation parameters from speech signals and can generalise to unseen speakers and diverse simulated environments. The results also indicate that the use of multiple branches disentangles the embedding space from misalignments between input features and subtasks.
Authors:
Thoidis, Iordanis; Vryzas, Nikolaos; Vrysis, Lazaros; Kotsakis, Rigas; Kalliris, George; Dimoulas, Charalampos
Affiliation:
Aristotle University of Thessaloniki, Greece
AES Convention:
152 (May 2022)
Paper Number:
10593
Publication Date:
May 2, 2022
Subject:
Machine Learning / Artificial Intelligence
Click to purchase paper as a non-member or you can login as an AES member to see more options.
No AES members have commented on this paper yet.
To be notified of new comments on this paper you can subscribe to this RSS feed. Forum users should login to see additional options.
If you are not yet an AES member and have something important to say about this paper then we urge you to join the AES today and make your voice heard. You can join online today by clicking here.