In This Section
Clean Audio for TV broadcast: An Object-Based Approach for Hearing-Impaired Viewers - April 2015
Audibility of a CD-Standard A/DA/A Loop Inserted into High-Resolution Audio Playback - September 2007
Sound Board: Food for Thought, Aesthetics in Orchestra Recording - April 2015
AES Conference Papers Forum
GPU-Based WFS Systems with Mobile Virtual Sound Sources and Room Compensation
Wave Field Synthesis (WFS) is a spatial audio reproduction system that provides an accurate spatial sound field in a wide area. This sound field is rendered through a high number of loudspeakers to emulate virtual sound sources. WFS systems require high computational capacity since they involve multiple loudspeakers and multiple virtual sources. Furthermore improvements of the spatial audio perception imply even higher processing capacity, mainly to avoid artifacts when the virtual sources move, and compensate the room e ects at certain control points within the listening area. Graphics Processing Units (GPUs) are well-known for their potential in highly parallel data processing. In this paper, we propose a GPU implementation that yields maximum parallelism by adapting the required computations to the di erent GPU architectures (Tesla, Fermi and Kepler).
No AES members have commented on this paper yet.
Subscribe to this discussion
Start a discussion!
If you are not yet an AES member and have something important to say about this paper then we urge you to join the AES today and make your voice heard. You can join online today by clicking here.