This paper proposes a new audio rendering method that tries to preserve the sound inputs to both ears instead of the sound direction. It uses a conversion matrix that converts the original sound signal into the converted sound signal with the different number of channels. The least squares method optimizes the matrix so as to minimize the difference between the input signals to both ears by the original signal and those by the rendered signals. To calculate the error function, the method uses the Head Related Impulse Responses. Two rendering experiments were conducted to evaluate the method. In the first experiment, 22 channel signals of 22.2 multichannel without two LFE channels were rendered into three dimensional 8-channel signals by the conventional directional-based method and the new method. The result showed that the new method could preserve the diffuseness of sound better than the conventional method. In the second experiment, the 22 channel signals were converted into 2-channel signals by the conventional downmix method and the new method. The evaluation result based on the cross correlation coefficient showed that there were not so many differences between the downmix method and the new method. However, the informal listening test showed that the new method might preserve the diffuseness of sound better than the downmix method.
Affiliation: University of Toyama, Toyama, Japan
AES Convention: 146 (March 2019) Paper Number: 10196
Publication Date: March 10, 2019
Subject: Poster Session 3
No AES members have commented on this paper yet.
If you are not yet an AES member and have something important to say about this paper then we urge you to join the AES today and make your voice heard. You can join online today by clicking here.