We use cookies on this website. By continuing to use this site without changing your cookie settings, you agree that you are happy to accept our cookies and for us to access these on your device. Find out more about how we use cookies and how to change your cookie settings.

Brain picks out salient sounds from background noise by tracking frequency and time

23 July 2013

New research reveals how our brains are able to pick out important sounds from the noisy world around us. The findings, published online today in the journal ‘eLife’, could lead to new diagnostic tests for hearing disorders.

Our ears can effortlessly pick out the sounds we need to hear from a noisy environment - hearing our mobile phone ringtone in the middle of the Notting Hill Carnival, for example - but how our brains process this information (the so-called 'cocktail party problem') has been a longstanding research question in hearing science.

Researchers have previously investigated this using simple sounds such as two tones of different pitches, but now researchers at UCL and Newcastle University have used complicated sounds that are more representative of those we hear in real life. The team used 'machine-like beeps' that overlap in both frequency and time to recreate a busy sound environment and obtain new insights into how the brain solves this problem.

In the study, groups of volunteers were asked to identify target sounds from within this noisy background in a series of experiments.

Sundeep Teki, a PhD student from the Wellcome Trust Centre for Neuroimaging at UCL and joint first author of the study, said: "Participants were able to detect complex target sounds from the background noise, even when the target sounds were delivered at a faster rate or there was a loud disruptive noise between them."

Dr Maria Chait, a senior lecturer at UCL Ear Institute and joint first author on the study, adds: "Previous models based on simple tones suggest that people differentiate sounds based on differences in frequency, or pitch. Our findings show that time is also an important factor, with sounds grouped as belonging to one object by virtue of being correlated in time."

Professor Tim Griffiths, Professor of Cognitive Neurology at Newcastle University and lead researcher on the study, said: "Many hearing disorders are characterised by the loss of ability to detect speech in noisy environments. Disorders like this that are caused by problems with how the brain interprets sound information, rather than physical damage to the ear and hearing machinery, remain poorly understood.

"These findings inform us about a fundamental brain mechanism for detecting sound patterns and identifies a process that can go wrong in hearing disorders. We now have an opportunity to create better tests for these types of hearing problems."

The research was funded by the Wellcome Trust and Deafness Research UK.

Image: Commuters in Liverpool Street Station. Credit: John Seb Barber on Flickr.

Contact

Jen Middleton
Senior Media Officer, Wellcome Trust
T
020 7611 7262
E
j.middleton@wellcome.ac.uk

Notes to editors

Reference
Teki S et al. Segregation of complex acoustic scenes based on temporal coherence. eLife 2013 (epub).

About the Wellcome Trust
The Wellcome Trust is a global charitable foundation dedicated to achieving extraordinary improvements in human and animal health. It supports the brightest minds in biomedical research and the medical humanities. The Trust's breadth of support includes public engagement, education and the application of research to improve health. It is independent of both political and commercial interests.

Wellcome Trust, Gibbs Building, 215 Euston Road, London NW1 2BE, UK T:+44 (0)20 7611 8888