Disguise or masking, which is the equivalent of pixelation in video, can be applied in an audio setting. We examine the application of semi-automatic speaker recognition to aid segmentation of audio intended for disguise or masking. In this work, we have proposed an innovative application of semi-automatic speaker recognition to automate audio segmentation in the process of disguising or masking the identity of vulnerable witnesses or undercover officers in police interviews. In this study, the accuracy of the semi-automatic speaker segmentation process was measured against baseline references as annotated by a human audio expert. Metrics were also established for estimating the accuracy of segmentation, and these include overlap errors, border accuracy, and feathered border accuracy. These were measured against analysis block sizes, number of Gaussian components, and training data length. Our experimentation has shown low error rates for speaker separation, and has also shown that accuracy metrics need to take transition boundaries between speakers into account. We have also considered the influence of the content of the training utterances and evaluated the influence of noise on performance. This approach was tested against real-case data used by the London Metropolitan Police.
Alexander, Anil; Forth, Oscar; How, Robin
Affiliations: GriffComm Ltd., Oxford, UK; Metropolitan Police, London, UK(See document for exact affiliation information.)
AES Conference: 39th International Conference: Audio Forensics: Practices and Challenges (June 2010)
Paper Number: 6-1
Publication Date: June 17, 2010
Subject: Speech and Forensics - Voice Identification
No AES members have commented on this paper yet.
If you are not yet an AES member and have something important to say about this paper then we urge you to join the AES today and make your voice heard. You can join online today by clicking here.