In this paper we present an automatic vocal percussion transcription system aimed to be an alternative to touchscreen input for drum and percussion programming. The objective of the system is to simplify the workflow of the user by letting him create percussive tracks made up of different samples triggered by his own voice without the need of any demanding skill by creating a system tailored to his specific needs. The system consists of four stages: event detection, feature extraction, and classification. We are employing small user-generated databases to adapt to particular vocalizations while avoiding overfitting and maintaining computational complexity as low as possible.
Authors:
Sánchez-Hevia, Héctor A.; Llerena-Aguilar, Cosme; Ramos-Auñón, Guillermo; Gil-Pita, Roberto
Affiliation:
University of Alcala, Alcalá de Henares, Madrid, Spain
AES Convention:
138 (May 2015)
Paper Number:
9352
Publication Date:
May 6, 2015
Subject:
Semantic Audio
Click to purchase paper as a non-member or you can login as an AES member to see more options.
No AES members have commented on this paper yet.
To be notified of new comments on this paper you can subscribe to this RSS feed. Forum users should login to see additional options.
If you are not yet an AES member and have something important to say about this paper then we urge you to join the AES today and make your voice heard. You can join online today by clicking here.