Directional audio coding (DirAC) is a parametric approach for the analysis and reproduction of spatial sound. The DirAC parameters, namely direction-of-arrival and diffuseness of sound can be further exploited in modern teleconferencing systems. Based on the directional parameters, we can control a video camera to automatically steer on the active talker. In order to provide consistency between the visual and acoustical cues, the virtual recording position should match the visual movement. In this paper, we present an approach for an acoustical zoom, which provides audio rendering that follows the movement of the visual scene. The algorithm does not rely on a priori information regarding the sound reproduction system as it operates directly in the DirAC parameter domain.
Authors:
Schultz-Amling, Richard; Kuech, Fabian; Thiergart, Oliver; Kallinger, Markus
Affiliation:
Fraunhofer Institute for Integrated Circuits IIS, Erlangen, Germany
AES Convention:
128 (May 2010)
Paper Number:
8120
Publication Date:
May 1, 2010
Subject:
Multichannel and Spatial Audio
Click to purchase paper as a non-member or you can login as an AES member to see more options.
No AES members have commented on this paper yet.
To be notified of new comments on this paper you can subscribe to this RSS feed. Forum users should login to see additional options.
If you are not yet an AES member and have something important to say about this paper then we urge you to join the AES today and make your voice heard. You can join online today by clicking here.