In mixed reality (MR) applications, digital audio objects are rendered via an acoustically transparent playback system to blend with the physical surroundings of the listener. This requires a binaural simulation process that perceptually matches the reverberation properties of the local environment, so that virtual sounds are not distinguishable from real sounds emitted around the listener. In this paper we propose an acoustic scene programming model that allows pre-authoring the behaviors and trajectories of a set of sound sources in a MR audio experience, while deferring to rendering time the specification of the reverberation properties of the enclosing room.
Authors:
Audfray, Rémi; Jot, Jean-Marc; Dicker, Sam
Affiliation:
Magic Leap, Inc., San Francisco, CA, USA
AES Convention:
145 (October 2018)
eBrief:477
Publication Date:
October 7, 2018
Subject:
Spatial Audio
Download Now (496 KB)
This paper is Open Access which means you can download it for free.
No AES members have commented on this paper yet.
To be notified of new comments on this paper you can
subscribe to this RSS feed.
Forum users should login to see additional options.
If you are not yet an AES member and have something important to say about this paper then we urge you to join the AES today and make your voice heard. You can join online today by clicking here.