Skip to main content
  • Poster presentation
  • Open access
  • Published:

Spiking neural network model of cortical auditory source segregation

Humans have the remarkable ability to tune into a particular voice even in loud noisy environments. The neural underpinnings of this amazing perceptual phenomenon are not yet fully understood. Recent EcoG [1] and MEG studies [2] have established that the neural representation of the attended speaker’s speech is much stronger than the unattended (distractor) speech when human subjects are asked to pay attention to a target speaker in a mixture of speech. How the brain sieves through the mixture waveform to enhance the target speaker’s speech and attenuate the background acoustic scene is still being investigated. In this work, we propose a spiking neural network architecture based on the theory of temporal coherence [3] to achieve auditory source segregation.

Our model does not require training on the background noise or prior exposure to the target speech. Along with using bottom-up spectro-temporal features and pitch features, the model can also accommodate top-down attentional mechanisms to generate segregated phase locked neural representations to target speaker’s speech envelope. The model comprises of a feature extraction stage followed by clustering stage. The feature extraction stage mimics the auditory pathway starting from a cochlear representation followed by a multi-resolution analysis of the cochlear output using a bank of band-pass filters (cortical stage), to provide a rich timbre representation. Dominant pitch tracks are extracted from the sound mixture and processed through the same set of band-pass filters as the timbre channels. The output of the feature extraction stage comprising of the pitch and timbre channels are transduced into a spike-based representation using leaky integrate and fire neurons with time constants tuned to the bandwidth of the multi-resolution band-pass filters. The clustering stage comprises of a bank of coincidence detector neurons. Using the pitch signals as anchors the coincidence detector neurons can segregate the two sources from the mixture timbre representation. Thus, the output of the coincidence detector neurons comprises only of responses phase locked to the envelope of a single source.

This model does not require any weight learning, is unsupervised and can segregate sources online. Previous studies on correlation based sound segregation employed network of neurons with intrinsic oscillator dynamics [4]. In this work, clustering of features belonging to a single source is driven only by the temporal coherence of spectro-temporal features of the given source. This spike-based representation provides an easy mechanism to group coherent features, which otherwise would require computationally expensive numerical routines for online, adaptive principal components analysis. Future work is aimed at reconstructing the speech waveform from the segregated spike trains.

References

  1. Mesgarani N, Chang E: Selective cortical representation of attended speaker in multi-talker speech perception. Nature. 2012, 485: 233-236. 10.1038/nature11020.

    Article  CAS  PubMed  Google Scholar 

  2. Ding N, Simon JZ: Emergence of neural encoding of auditory objects while listening to competing speakers. Proceedings of the National Academy of Sciences. 2012, 109: 11854-11859. 10.1073/pnas.1205381109.

    Article  CAS  Google Scholar 

  3. Shamma S, Elhilali M, Micheyl C: Temporal coherence and attention in auditory scene analysis. Trends in Neurosciences. 2011, 34: 114-123. 10.1016/j.tins.2010.11.002.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  4. Wang D, Brown G: Separation of speech from interfering sounds based on oscillatory correlation. IEEE Transactions on Neural Networks. 1999, 10: 684-697. 10.1109/72.761727.

    Article  CAS  PubMed  Google Scholar 

Download references

Acknowledgements

The spike-based architecture for source segregation was developed when the first author was interning with Qualcomm Research, San Diego.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Lakshmi Krishnan.

Rights and permissions

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Krishnan, L., Campos, M. & Shamma, S. Spiking neural network model of cortical auditory source segregation. BMC Neurosci 15 (Suppl 1), P50 (2014). https://0-doi-org.brum.beds.ac.uk/10.1186/1471-2202-15-S1-P50

Download citation

  • Published:

  • DOI: https://0-doi-org.brum.beds.ac.uk/10.1186/1471-2202-15-S1-P50

Keywords