When an audio recording is used as evidence in litigation and forensic investigations, it needs to be checked thoroughly for authenticity and integrity in order to be admissible, compelling, and decisive evidence in a court of law. An audio recording can be subject to tampering attacks with easy-to-use editing and signal processing tools, thereby undermining its legal value. Artifacts embedded in an audio recording can provide valuable clues about the acoustic environment in` which the audio was recorded and allow for the detection of tampering. This paper presents findings of two parallel methodologies: (1) where the features are extracted from the room impulse response and (2) where features are extracted directly from the reverberated recordings. These methods focus on extracting parameters from audio recordings that helped distinguish different auditory scenes. Experiments employing an exhaustive set of machine learning classifiers along with different acoustic features were conducted for the classification of auditory environments. A comparative analysis has been carried out to assess the performance of each classifier and relative performance impact of each feature set in terms of the accuracy of classification. A two-layer Artificial Neural Network (ANN) provided an accuracy of 98.7% using room impulse responses and an accuracy of 99.5% when the reverberated audio recordings were trained.
Patole, Rashmika; Rege, Priti
Affiliation: College of Engineering, Pune, India
JAES Volume 67 Issue 12 pp. 939-952; December 2019
Publication Date: December 30, 2019
No AES members have commented on this paper yet.
If you are not yet an AES member and have something important to say about this paper then we urge you to join the AES today and make your voice heard. You can join online today by clicking here.