We propose a new method to automatically determine key signature changes. In automatic music transcription, sections in distantly related keys may lead to music scores that are hard to read due to a high number of notated accidentals. The problem of key change is commonly addressed by finding the correct local key out of the 24 major and minor keys. However, to provide the best matching key signature, choosing the right mode (major or minor) is not necessary and thus, we only estimate the local underlying diatonic scale. After extracting chroma features and a beat grid from the audio data, we calculate local probabilities for the different diatonic scales. For this purpose, we present a multiplicative procedure that shows promising results for visualizing complex tonal structures. From the obtained diatonic scale estimates, we identify candidates for key signature changes. By clustering similar segments and applying minimum segment length constraints, we get the tonal segmentation. We test our method on a dataset containing 30 hand-annotated pop songs. To evaluate our results, we calculate scores based on the number of frames correctly annotated, as well as segment border F-measures and perform a cross-validation study. Our rule-based method yields up to 90% class accuracy and up to 70% F-measure score for segment borders. These results are promising and qualify the approach to be applied for automatic music transcription.
Weiss, Christof; Cano, Estefania; Lukashevich, Hanna
Affiliation: Fraunhofer Institute for Digital Media Ttechnology, Ilmenau, Germany
AES Conference: 53rd International Conference: Semantic Audio (January 2014)
Paper Number: P1-2
Publication Date: January 27, 2014
Subject: Music Informatics and Retrieval
No AES members have commented on this paper yet.
If you are not yet an AES member and have something important to say about this paper then we urge you to join the AES today and make your voice heard. You can join online today by clicking here.