Computational Auditory Scene Analysis (CASA) is typically achieved by using statistical models that have been trained offline on available data. Their performance relies heavily on the assumption that the process generating the data along with the recording conditions are stationary over time. Nowadays, the focus of CASA is moving from structured, well-defined scenarios to unrestricted scenes with realistic characteristics where the stationarity assumption might not be true. Therefore, there is a high demand for methodologies and tools dealing with a series of problems tightly coupled with such nonstationary conditions, such as changes in the recording conditions, reverberant effects, etc. This paper formulates these obstacles under the concept drift framework and explores two fundamental adaptation approaches: active and passive. The overall aim is to learn online the statistical properties of the evolving data distribution and incorporate them into the recognition mechanism for boosting its performance. The proposed CASA system encompasses a concept drift detector and an online adaptation module. The proposed framework was evaluated in the auditory analysis of three environments (office, meeting room, and lecture hall) with diverse characteristics (dimensions, reverberation times, etc.) The results are encouraging in terms of classification rate, false positive rate, false negative rate, and detection delay.
Affiliation: Department of Computer Science, University of Milan Via Celoria, Milan, Italy
JAES Volume 67 Issue 10 pp. 772-781; October 2019
Publication Date: October 25, 2019
No AES members have commented on this paper yet.
If you are not yet an AES member and have something important to say about this paper then we urge you to join the AES today and make your voice heard. You can join online today by clicking here.