A method is presented for high-quality recording of voice and acoustic instruments in loudspeaker-generated virtual acoustics. Auralization systems typically employ close-mic'ing to avoid feedback, while classical recording methods prefer high-quality room microphones to capture the instruments integrated with the space. Popular music production records dry tracks, and applies reverberation after primary edits are complete. Here a hybrid approach is taken, using close mics to produce real-time, loudspeaker-projected virtual acoustics, and room microphones to capture a balanced, natural sound. The known loudspeaker signals are then used to cancel the virtual acoustics from the room microphone tracks, providing a set of relatively dry tracks for use in editing and post-production. Example recordings of Byzantine chant in a virtual Hagia Sophia are described.
Authors:
Abel, Jonathan S.; Canfield-Dafilou, Elliot K.
Affiliations:
Stanford University, Stanford, CA, USA; Center for Computer Research in Music and Acosutics (CCRMA), Stanford University, Stanford, CA, USA(See document for exact affiliation information.)
AES Convention:
143 (October 2017)
Paper Number:
9854
Publication Date:
October 8, 2017
Subject:
Recording and Production
Click to purchase paper as a non-member or you can login as an AES member to see more options.
No AES members have commented on this paper yet.
To be notified of new comments on this paper you can subscribe to this RSS feed. Forum users should login to see additional options.
If you are not yet an AES member and have something important to say about this paper then we urge you to join the AES today and make your voice heard. You can join online today by clicking here.