Recent developments in AR / VR applications have brought a renewed focus on efficient and scalable real-time HRTF renderers to alleviate compute constraints when spatializing many sound sources at once. To efficiently achieve a reasonable approximation of the full-sphere, the HRTF dataset is often linearly decomposed into a predetermined number of basis filters via methods such as Ambisonics, VBAP, or PCA. This paper proposes a novel HRTF renderer and decomposition technique that, when compared to previous methods, allows for greater accuracy of the HRTF approximation for an equivalent compute cost. This is achieved through a multi-layered optimization network architecture that minimizes a perceptually motivated error function to derive the basis filters. We will demonstrate the numerical accuracy of our technique as well as provide listening test results comparing our method to other linear decomposition methods of relative computational cost using both our internal and the publicly available SADIE HRTF datasets.
Authors:
Marchan, Mick; Allen, Andrew
Affiliation:
Microsoft, Redmond, WA, USA
AES Conference:
2022 AES International Conference on Audio for Virtual and Augmented Reality (August 2022)
Paper Number:
9
Publication Date:
August 15, 2022
Subject:
Paper
Click to purchase paper as a non-member or you can login as an AES member to see more options.
No AES members have commented on this paper yet.
To be notified of new comments on this paper you can
subscribe to this RSS feed.
Forum users should login to see additional options.
If you are not yet an AES member and have something important to say about this paper then we urge you to join the AES today and make your voice heard. You can join online today by clicking here.