A theory and a system for capturing an audio scene and then rendering it remotely are developed and presented. The sound capture is performed with a spherical microphone array. The sound field at the location, and in a region of space in the neighborhood, of the array is deduced from the captured sound and represented using either spherical wave-functions or plane-wave expansions. The representation is then transmitted to a remote location for immediate rendering or stored for later reproduction. The sound renderer, coupled with the head tracker, reconstructs the acoustic field using individualized head-related transfer functions to preserve the perceptual spatial structure of the audio scene. Rigorous error bounds and a Nyquist-like sampling criterion for the representation of the sound field are presented and verified.
Davis, Larry S.; Duraiswami, Ramani; Grassi, Elena; Gumerov, Nail A.; Li, Zhiyun; Zotkin, Dmitry N.
Affiliation: Perceptual Interfaces and Reality Laboratory, University of Maryland
AES Convention: 119 (October 2005) Paper Number: 6540
Publication Date: October 1, 2005
Subject: Spatial Perception and Processing
No AES members have commented on this paper yet.
If you are not yet an AES member and have something important to say about this paper then we urge you to join the AES today and make your voice heard. You can join online today by clicking here.