This paper presents a multichannel audio coding scheme based on existing parametric spatial audio coding schemes, which aims at both data compression and quality improvement. It provides a dynamic quantization method of the spatial parameters relying on an instantaneous estimation of human auditory spatial resolution. Psychoacoustic experiments studying this resolution for complex sound scenes are reported. They consist of measures of azimuthal localization blur through minimum audible angles. Results mainly show that the spatial auditory resolution is likely to degrade as the scene gets more complex. A modeling of this phenomenon is also outlined.
Authors:
Daniel, Adrien; Nicol, Rozenn; McAdams, Stephen
Affiliations:
CIRMMT, Schulich School of Music, McGill University, Montreal, Quebec, Canada; Orange Labs - Lannion, Lannion, France(See document for exact affiliation information.)
AES Conference:
45th International Conference: Applications of Time-Frequency Processing in Audio (March 2012)
Paper Number:
5-1
Publication Date:
March 1, 2012
Subject:
Psychoacoustics and Hearing
Click to purchase paper as a non-member or you can login as an AES member to see more options.
No AES members have commented on this paper yet.
To be notified of new comments on this paper you can subscribe to this RSS feed. Forum users should login to see additional options.
If you are not yet an AES member and have something important to say about this paper then we urge you to join the AES today and make your voice heard. You can join online today by clicking here.