Recently, people who prefer to consume media contents via over the top (OTT) platform, such as YouTube, Netflix etc., rather than a conventional broadcasting get increased more and more. To deliver an immersive audio experience to them more effectively, we propose a unified framework for AI-based 3D immersive audio codec. In this framework, to maximize the original immersiveness even at a down-mixed audio, while enabling to precisely reproduce the original 3D audio from the down-mixed audio, content-adaptive dynamic down-mixing and up-mixing scheme is newly proposed. The experimental results show that the proposed framework can render more improved down-mixed audio compared to the conventional method as well as successfully reproduce the original 3D audio.
Authors:
Nam, Woo Hyun; Lee, Tammy; Ko, Sang Chul; Son, Yoonjae; Chung, Hyun Kwon; Kim, Kyung-Rae; Kim, Jungkyu; Hwang, Sunghee; Lee, Kyunggeun
Affiliation:
Samsung Research, Samsung Electronics, Seoul, Republic of Korea
AES Convention:
151 (October 2021)
Paper Number:
10525
Publication Date:
October 13, 2021
Subject:
Multichannel and spatial audio processing and applications
Click to purchase paper as a non-member or you can login as an AES member to see more options.
No AES members have commented on this paper yet.
To be notified of new comments on this paper you can
subscribe to this RSS feed.
Forum users should login to see additional options.
If you are not yet an AES member and have something important to say about this paper then we urge you to join the AES today and make your voice heard. You can join online today by clicking here.