Identifying the signal in the noise: a systems approach for examining invariance in auditory cortex

Lead Research Organisation: University College London
Department Name: Ear Institute

Abstract

We are able to recognize and understand speech across many different speakers, voice pitches and listening conditions. However, the acoustic waveform of a sound (e.g. the vowel 'ae') will vary considerably depending on the individual speaker, and the 'ae' may be embedded in a cacophony of other, background sounds in our often noisy environments. Despite this, we have no difficulty recognizing an 'ae' as an 'ae', suggesting that the brain is capable of forming a representation of the vowel sound which is invariant to these 'nuisance' variables. For vowel sounds, the timbre, or vowel identity, is determined by the spectral envelope. Filtering by the mouth, lips and tongue results in energy peaks, or 'formants' in the spectrum, and it is the location of these formants which differentiates vowel sounds from one another. Thus, the fact that we are able to discriminate 'ae' from 'ih' irrespective of the gender, age or accent of a speaker suggests that we are able to form an invariant representation of the formant relations independently of the fundamental frequency, room reverberations, or spatial location in both quiet and noisy conditions. The aim of this research program is to discover where and how such invariant representations arise in the central auditory system and how these representations are maintained in noisy environments. Forming invariant representations is one of the greatest challenges for sensory systems, and understanding where and how such representations are read out is crucial for the design of any neuroprosthetic device. Our research uses ferrets as their hearing range spans a very similar range of frequencies to ours. Moreover, ferret vocalizations share many similarities with human vowel sounds. Ferrets rapidly learn to discriminate vowel sounds and we are able to record the activity of their nerve cells whilst they perform such listening tasks. By probing the circumstances under which the ferret is able to discriminate vowel sounds, and measuring the neural activity, we can look for where in the auditory brain invariant vowel representation might occur. The second part of this project involves reversibly silencing individual brain areas by cooling them. The principle of this technique is much the same was as using an ice pack to cool pain neurons in a bruised piece of skin. Small 'cryoloops' are implanted above auditory cortex in trained animals.This technique allows us to test whether particular brain areas are causally involved in vowel discrimination. The final part of this project investigates the role of visual information in auditory perception. It is well known that seeing a persons mouth movements while they talk to you enhances your ability to understand them - especially if you are listening in a very noisy room. When trying to pick out a quiet sound in a noisy background knowing when the sound is likely to occur also enhances your ability to correctly identify it. It has recently been shown that visual information is integrated into the very earliest auditory cortical areas. However, quite how this visual information shapes our auditory perception is unknown. The work in this proposal seeks to examine how visual information helps a trained animal to identify vowel sounds more accurately, whilst simultaneously examining how the visual stimulus influences the behaviour of neurons in auditory cortex. Inappropriate integration of auditory and visual information is postulated to underlie schizophrenic symptoms and understanding how informative visual stimuli influence auditory cortical activity will provide valuable insight into how sensory integration occurs in the healthy brain.Hearing impaired individuals most frequently suffer from an inability to effectively identify speech in noisy environments. Understanding how neurons are able to represent vowel identity robustly across a variety of listening conditions and noise environments will enhance hearing aid and cochlear implant design.

Technical Summary

Failure to understand speech in a noisy environment is one of the principal complaints of the hearing impaired. Our remarkable ability to recognize and understand speech across many different speakers, voice pitches and listening conditions likely depends on the auditory brain extracting those acoustic cues which provide reliable information about a particular sound feature in order to form a neural representation which robustly identifies the stimulus regardless of task-irrelevant 'nuisance' variables. This proposal employs a systems neuroscience approach to examine how auditory cortex forms invariant neural representations of vowel identity enabling a listener to differentiate vowels such as /ae/ from /ih/ irrespective of the voice pitch or location in space, and how this is maintained despite changes in the background noise environment. Three specific questions will be addressed: (1) How (with what neural code) and where (in which cortical field) do neurons support invariant perception of vowel identity? (2) Which cortical fields are necessary for an animal to perform invariant vowel recognition? (3) What role does visual information play in helping us to 'hear better'? By simultaneously measuring spiking activity, local field potentials and neural oscillations in trained ferrets performing a vowel identification task we will examine the incidence and location of invariant vowel timbre encoding. We will explore the neural codes which might support invariant coding and use reversible inactivation techniques examining causal relationships between the activity in specific cortical fields and vowel discrimination behaviour. Lastly we will examine how visual information is integrated within auditory cortex in order to aid listening in difficult conditions. Whilst multisensory integration has been documented in early sensory cortices very few studies have sought to correlate physiological measures with simultaneous assessment of any multisensory behavioural advantage.

Planned Impact

The proposed work is a fundamental neuroscience research project and will produce key insights into the functional organisation of mammalian sensory pathways and the processing of sounds by biological systems. Discriminating speech sounds in a noisy environment is one of the principle complaints of hearing impaired listeners. Understanding the neural processing which underlies this ability in normal hearing brains will offer key insights into how to better design signal processors in cochlear implants and hearing aids. Similar advantages will be afforded to communication technologies; we need only consider the very substantial shortcomings of even the most artificial speech recognition systems to be reminded of the remarkable sophistication of the auditory system. Collaboration with ENT surgeons and audiologists both within and beyond Oxford University ensures that our work maintains a clinical focus. Dr Hartley, a clinician-scientist has recently developed a ferret cochlear implant model meaning that the neural processing insights we gain from our efforts to relate perception to neural firing can be used to generate testable hypothesis which can be implemented within this animal model. We have established collaborations within Oxford with ENT clinicians. We maintain active collaborations with the growing number of sensory neuroscience groups, both in the UK and USA, who use ferrets as an animal model. Through regular dialogue we will continue to share new data, techniques and methods to the benefit of all. An example of such collaborative endeavour is the ferret brain atlas which is being developed by neuroanatomist Dr Sussane Radtke-Schuller (based in Munich), in collaboration with the University of Maryland based group of Dr Shihab Shamma, and the Oxford group. This atlas is the first of its kind for this species and will provide a free online resource containing cytoarchitectonic data and high resolution structural MRI scans which will be available to assist the increasing number of scientists who are working with ferrets. We increase the impact of our work by participate in the Deafness Research UK's public outreach programs. In the past year we have had a number of school work experience students, and a Nuffield bursary student spend time in the lab in an attempt to encourage more school leavers to consider pursuing a career in biomedical science. We maintain an uptodate website which details our most recent work as well as our research goals and includes routes through which members of the public can contact us. The research in this proposal addresses fundamental neuroscience issues with little immediate commercial exploitation potential. Its benefits will be in enhancing our scant knowledge of how the healthy brain operates to process sounds. This knowledge will be, through publication and dissemination at international meetings, made available to engineers and others who will be able to apply what we learn about neural coding to improving the signal processing capabilities of cochlear implants, hearing aids and telecommunication devices. Whilst cochlear implants have been tremendously successful they are only an option for those individuals with an intact auditory nerve. A knowledge of the neural coding mechanisms within auditory cortex will guide the stimulation strategies and design of neural prosthesis which targets auditory centres in the brainstem or midbrain.
 
Description We are able to recognize and understand speech across many different speakers, voice pitches and listening conditions. However, the acoustic waveform of a sound (e.g. the vowel /a/) will vary considerably depending on the individual speaker, and the /a/; may be embedded in a cacophony of other, background sounds in our often noisy environments. Despite this, we have no difficulty recognizing an /a/; as an /a/, suggesting that the brain is capable of forming a representation of the vowel sound which is invariant to these nuisance variables. Thus, the fact that we are able to discriminate /a/ from /e/; irrespective of the gender, age or accent of a speaker suggests that we are able to form an invariant representation of the formant relations independently of the fundamental frequency, room reverberations, or spatial location in both quiet and noisy conditions. The aim of this research program was to discover where and how such invariant representations arise in the central auditory system and how these representations are maintained in noisy environments. We have recorded neural activity during sound discrimination and used machine learning approaches to demonstrate how neurons are capable of achieving perceptual invariance across voice pitch and in different noise backgrounds. We have further documented how auditory cortical neurons are shaped by experience - both short term (for example when passively listening as opposed to actively discriminating) and long term (trained versus naive subjects). These findings were published in 2018 in Nature Communications. Our second paper key is under review and available on BioRXiv and demonstrates a casual role for auditory cortex in sound processing in noise, as well as determining how single neurons and neural populations support (and equally importantly, fail to support) listening in noise. The scope of this grant was extended (in collaboration with a PhD student) to study invariance of spatial processing - we demonstrated that a sub-population of auditory cortical neurons represent the spatial location of a sound source in a manner that is invariant to the precise acoustic cues (Wood, Town and Bizley, Nature Communications, 2019).
Exploitation Route Our findings have been presented at a variety of national and international meetings. We have published two review papers and have a paper detailing some behavioural work undergoing a second revision. Our neural findings have been published or are under review..

Our software and methods have been shared with a number of other researchers. In addition the methods that we developed for freely moving recordings have led to me being able to assisted another research group in setting up a ferret lab as an alternative to primate work.

We hope that our findings will provide the foundation for work exploring how the auditory brain makes sense of complicated sound scenes - such as listening to conversation in a busy restaurant. Our data provide insight into how the brain achieves this challenge and may in the future provide biologically inspired solutions to computer listening devices or signal processing devices in hearing prosthetics.
Sectors Digital/Communication/Information Technologies (including Software),Healthcare

 
Description Defeating Deafness PhD studentship
Amount £65,000 (GBP)
Organisation Action on Hearing Loss 
Sector Charity/Non Profit
Country United Kingdom
Start 09/2013 
End 04/2017
 
Description Human Frontiers Science Foundation
Amount $375,000 (USD)
Funding ID RGY0068 
Organisation Human Frontier Science Program (HFSP) 
Sector Charity/Non Profit
Country France
Start 12/2014 
End 11/2017
 
Description International Award
Amount £12,000 (GBP)
Organisation The Royal Society 
Sector Charity/Non Profit
Country United Kingdom
Start 07/2012 
End 07/2015
 
Description Royal Society / Wellcome Trust Sir Henry Dale Fellowship
Amount £1,500,000 (GBP)
Funding ID WT098418RR 
Organisation Sir Henry Dale Fellowships 
Sector Charity/Non Profit
Country United Kingdom
Start 09/2013 
End 08/2018
 
Description Prof Adrian KC Lee, University of Washington, Seattle 
Organisation University of Washington
Country United States 
Sector Academic/University 
PI Contribution We initially collaborated on a psychophysics project to explore audiovisual binding and have sinced worked together to share code, data and ideas to develop the theoretical basis for these experiments. This collaboration begun during my time as a research fellow in Boston, and continued, funded through a Royal Society international award from 2012-2015.
Collaborator Contribution With our collaborators we designed the original psychophysical paradigm which was run in parallel in our two labs. We are performing animal studies while our collaborators perform imaging - in both cases using the same stimuli. This resulted in multiple publications (Maddox et al., eLife, 2015, Atilgan et al., Neuron, 2018, Bizley, Maddox, Lee, TINS 2016; Lee, Maddox, Bizley, 2019).
Impact Publications: Bizley, Shinn-Cunningham and Lee 2012, J.Neurosci. Maddox et al., 2015, eLife; Bizley et al, TINs, 2016; Atilgan et al., 2018, Neuron. Lee et al., SHAR 2019
Start Year 2010
 
Description UNC Chapel Hill, North Carolina 
Organisation University of North Carolina at Chapel Hill
Department Department of Psychiatry
Country United States 
Sector Academic/University 
PI Contribution We have a collaborative project which has been funded by HFSP and which commences this month.
Collaborator Contribution We have a collaborative project which has been funded by HFSP and which commences this month.
Impact None yet.
Start Year 2013
 
Title GoFerret 
Description Software to interface with TDT equipment to control behavioural testing apparatus. This was developed by the post-doc and PI named on this grant (Town and Bizley) and has been distributed to at least two other laboratories. 
Type Of Technology Software 
Year Produced 2013 
Impact No actual Impacts realised to date 
 
Title Neural analysis software 
Description Analysis suite (written by Dr Stephen Town) for collection and analysis of neural data. Provided to 4 research groups on request. 
Type Of Technology Software 
Year Produced 2014 
Impact Provided to 4 research groups on request. 
 
Description Auditory Scene Analysis Workshop HWK, Delmenhorst 
Form Of Engagement Activity A press release, press conference or response to a media enquiry/interview
Part Of Official Scheme? No
Geographic Reach International
Primary Audience Other academic audiences (collaborators, peers etc.)
Results and Impact Discussion and continuing collaboration
Year(s) Of Engagement Activity
 
Description BBC news online associated with BBC2's "Trust me I'm a Doctor" episode. 
Form Of Engagement Activity A magazine, newsletter or online publication
Part Of Official Scheme? No
Geographic Reach International
Primary Audience Public/other audiences
Results and Impact Interview for a news article related to a BBC documentary.
Year(s) Of Engagement Activity 2018
URL http://www.bbc.co.uk/news/health-42620387
 
Description BBC2 Trust me I'm a Doctor appearance 
Form Of Engagement Activity A broadcast e.g. TV/radio/film/podcast (other than news/press)
Part Of Official Scheme? No
Geographic Reach International
Primary Audience Public/other audiences
Results and Impact I was invited onto the program as a neuroscientist to provide an explanation for how lip reading can help deaf and hearing impaired listeners to listen more effectively.
Year(s) Of Engagement Activity 2018
URL https://www.bbc.co.uk/iplayer/episode/b09m62j5/trust-me-im-a-doctor-series-7-episode-2
 
Description Interview for national news paper (The Times) 
Form Of Engagement Activity A press release, press conference or response to a media enquiry/interview
Part Of Official Scheme? No
Geographic Reach National
Primary Audience Public/other audiences
Results and Impact Article written for the Saturday Times about our recent Neuron paper which demonstrated that visual inputs can help auditory cortex process sound mixtures - and therefore potentially help listening in everyday noisy situations.
Year(s) Of Engagement Activity 2018
URL https://www.thetimes.co.uk/edition/news/read-my-lips-visual-clues-amplify-sound-3wgzdbq89
 
Description Science showoff 
Form Of Engagement Activity Participation in an activity, workshop or similar
Part Of Official Scheme? Yes
Geographic Reach Regional
Primary Audience Public/other audiences
Results and Impact Talk generated lots of interest and discussion afterwards
Year(s) Of Engagement Activity 2014
 
Description SoapBox Science 
Form Of Engagement Activity Participation in an activity, workshop or similar
Part Of Official Scheme? Yes
Geographic Reach Local
Primary Audience Public/other audiences
Results and Impact Many people (adults and children) listened for 20 mins or more.

People seemed interested and to have appreciated new things about hearing and the brain
Year(s) Of Engagement Activity 2014
URL http://soapboxscience.org.uk
 
Description TEDx talk 
Form Of Engagement Activity A talk or presentation
Part Of Official Scheme? Yes
Geographic Reach International
Primary Audience Public/other audiences
Results and Impact Lots of discussion, >1000 views on youtube
Year(s) Of Engagement Activity 2014
URL http://www.youtube.com/watch?v=GWUVRjtlg1Y