In this paper, we propose a system for synthesizing the human singing voice and the musical subtleties that accompany it. The system, LYRICOS, employs a concatenation-based text-to-speech method to synthesize arbitrary lyrics in a given language. Using information contained in a regular MIDI file, the system chooses units, represented as sinusoidal waveform model parameters, from an inventory of data collected from a professional singer, and concatenates these to form arbitrary lyrical phrases. Standard MIDI messages control parameters for the addition of vibrato, spectral tilt, and dynamic musical expression, resulting in a very natural-sounding singing voice.
Authors:
Macon, Michael; Jensen-Link, Leslie; George, E. Bryan; Oliverio, James; Clements, Mark
Affiliations:
Oregon Graduate Institute of Science and Technology, Portland, OR ; Georgia Institute of Technology, Atlanta, GA ; Momentum Data Systems, Atlanta,GA ; Texas Instruments, Dallas,TX(See document for exact affiliation information.)
AES Convention:
103 (September 1997)
Paper Number:
4591
Publication Date:
September 1, 1997
Subject:
Electronic Music Synthesis and Control
Click to purchase paper as a non-member or you can login as an AES member to see more options.
No AES members have commented on this paper yet.
To be notified of new comments on this paper you can
subscribe to this RSS feed.
Forum users should login to see additional options.
If you are not yet an AES member and have something important to say about this paper then we urge you to join the AES today and make your voice heard. You can join online today by clicking here.