An Oscillatory Correlation Frame work for Computational Auditory Scene Analysis
Brown, Guy J., Wang, DeLiang L.
–Neural Information Processing Systems
A neural model is described which uses oscillatory correlation to segregate speech from interfering sound sources. The core of the model is a two-layer neural oscillator network. A sound stream is represented by a synchronized population of oscillators, and different streams are represented by desynchronized oscillator populations. The model has been evaluated using a corpus of speech mixed with interfering sounds, and produces an improvement in signal-to-noise ratio for every mixture. 1 Introduction Speech is seldom heard in isolation: usually, it is mixed with other environmental sounds. Hence, the auditory system must parse the acoustic mixture reaching the ears in order to retrieve a description of each sound source, a process termed auditory scene analysis (ASA) [2] . Conceptually, ASA may be regarded as a two-stage process.
Neural Information Processing Systems
Dec-31-2000