Community

AES Conference Papers Forum

Fusing Grouping Cues for Partials Using Artificial Neural Networks

Document Thumbnail

The primary stages of auditory stream separation are modelled here as a bottom-up organisation of primitive spectro-temporal regions into composites, which provide a succinct description of the sound environment in terms of a limited number of salient sound events. Many years of research on auditory streaming have determined a number of qualitative comparisons between physical attributes of sounds, or cues, that are well-correlated with the extent to which stream segregation or fusion of simple test stimuli occurs in listening tests. However, the relative importance of these cues is difficult to determine, especially in natural sound environments. This work represents some exploratory stages in using an artificial neural network to learn how to integrate multiple cues in nonlinear manner, for sound object formation. As a precursor to a more complex auditory front-end, a sinusoidal tracking algorithm was used to obtain the initial set of "spectro-temporal regions" or partial trajectories.

Authors:
Affiliation:
AES Conference:
Paper Number:
Publication Date:
Subject:

Click to purchase paper as a non-member or you can login as an AES member to see more options.

No AES members have commented on this paper yet.

Subscribe to this discussion

RSS Feed To be notified of new comments on this paper you can subscribe to this RSS feed. Forum users should login to see additional options.

Start a discussion!

If you would like to start a discussion about this paper and are an AES member then you can login here:
Username:
Password:

If you are not yet an AES member and have something important to say about this paper then we urge you to join the AES today and make your voice heard. You can join online today by clicking here.

AES - Audio Engineering Society