EnhanceMusic: Machine Learning Challenges to Revolutionise Music Listening for People with Hearing Loss
Lead Research Organisation:
University of Salford
Department Name: Sch of Science,Engineering & Environment
Abstract
Every culture has music. It brings people together and shapes society. Music affects how we feel, tapping into the pleasure circuits of the brain. In the UK each year, the core music industry contributes £3.5bn to the economy (UK Music 2012) with 30 million people attending concerts and festivals (UK Music 2017). Music listening is widespread in shops, movies, ceremonies, live gigs, on mobile phones, etc.
Music is important to health and wellbeing. As a 2017 report by the All-Party Parliamentary Group on Arts, Health & Wellbeing demonstrates, "The arts can help keep us well, aid our recovery and support longer lives better lived. The arts can help meet major challenges facing health and social care: ageing, long-term conditions, loneliness and mental health. The arts can help save money in the health service and social care."
1 on 6 people in the UK has a hearing loss, and this number will increase as the population ages (RNID). Poorer hearing makes music harder to appreciate. Picking out lyrics or melody lines is more difficult; the thrill of a musician creating a barely audible note is lost if the sound is actually inaudible, and music becomes duller as high frequencies disappear. This risks disengagement from music and the loss of the health and wellbeing benefits it creates.
We need to personalise music so it works better for those with a hearing loss. We will consider:
1. Processing and remixing mixing desk feeds for live events or multitrack recordings.
2. Processing of stereo recordings in the cloud or on consumer devices.
3. Processing of music as picked up by hearing aid microphones.
For (1) and (2), the music can be broadcast directly to a hearing aid or headphones for reproduction.
For (1), having access to separate tracks for each musical instrument gives greater control over how sounds are processed. This is timely with future Object-Based Audio formats allowing this approach.
(2) is needed because we consume much recorded music. It's more efficient and effective to pre-process music than rely on hearing aids to improve the sound, as this allows more sophisticated signal processing.
(3) is important because hearing aids are the solution for much live music. But, the AHRC Hearing Aids for Music project found that 67% of hearing-aid users had some difficulty listening to music with hearing aids. Hearing aid research has focussed mostly on speech with music listening being relatively overlooked.
Audio signal processing is a very active and fast-moving area of research, but typically fails to consider those with a hearing loss. The latest techniques in signal processing and machine learning could revolutionise music for those with a hearing impairment. To achieve this we need more researchers to consider hearing loss and this can be achieved through a series of signal processing challenges. Such competitions are a proven technique for accelerating research, including growing a collaborative community who apply their skills and knowledge to a problem area.
We will develop tools, databases and objective models needed to run the challenges. This will lower barriers that currently prevent many researchers from considering hearing loss. Data would include the results of listening tests into how real people perceive audio quality, along with a characterisation of each test subject's hearing ability, because the music processing needs to be personalised. We will develop new objective models to predict how people with a hearing loss perceive audio quality of music. Such data and tools will allow researchers to develop novel algorithms.
The scientific legacy will be new approaches for mixing and processing music for people with a hearing loss, a test-bed that readily allows further research, better understanding of the audio quality required for music, and more audio and machine learning researchers considering the hearing abilities of the whole population for music listening.
Music is important to health and wellbeing. As a 2017 report by the All-Party Parliamentary Group on Arts, Health & Wellbeing demonstrates, "The arts can help keep us well, aid our recovery and support longer lives better lived. The arts can help meet major challenges facing health and social care: ageing, long-term conditions, loneliness and mental health. The arts can help save money in the health service and social care."
1 on 6 people in the UK has a hearing loss, and this number will increase as the population ages (RNID). Poorer hearing makes music harder to appreciate. Picking out lyrics or melody lines is more difficult; the thrill of a musician creating a barely audible note is lost if the sound is actually inaudible, and music becomes duller as high frequencies disappear. This risks disengagement from music and the loss of the health and wellbeing benefits it creates.
We need to personalise music so it works better for those with a hearing loss. We will consider:
1. Processing and remixing mixing desk feeds for live events or multitrack recordings.
2. Processing of stereo recordings in the cloud or on consumer devices.
3. Processing of music as picked up by hearing aid microphones.
For (1) and (2), the music can be broadcast directly to a hearing aid or headphones for reproduction.
For (1), having access to separate tracks for each musical instrument gives greater control over how sounds are processed. This is timely with future Object-Based Audio formats allowing this approach.
(2) is needed because we consume much recorded music. It's more efficient and effective to pre-process music than rely on hearing aids to improve the sound, as this allows more sophisticated signal processing.
(3) is important because hearing aids are the solution for much live music. But, the AHRC Hearing Aids for Music project found that 67% of hearing-aid users had some difficulty listening to music with hearing aids. Hearing aid research has focussed mostly on speech with music listening being relatively overlooked.
Audio signal processing is a very active and fast-moving area of research, but typically fails to consider those with a hearing loss. The latest techniques in signal processing and machine learning could revolutionise music for those with a hearing impairment. To achieve this we need more researchers to consider hearing loss and this can be achieved through a series of signal processing challenges. Such competitions are a proven technique for accelerating research, including growing a collaborative community who apply their skills and knowledge to a problem area.
We will develop tools, databases and objective models needed to run the challenges. This will lower barriers that currently prevent many researchers from considering hearing loss. Data would include the results of listening tests into how real people perceive audio quality, along with a characterisation of each test subject's hearing ability, because the music processing needs to be personalised. We will develop new objective models to predict how people with a hearing loss perceive audio quality of music. Such data and tools will allow researchers to develop novel algorithms.
The scientific legacy will be new approaches for mixing and processing music for people with a hearing loss, a test-bed that readily allows further research, better understanding of the audio quality required for music, and more audio and machine learning researchers considering the hearing abilities of the whole population for music listening.
Organisations
- University of Salford (Lead Research Organisation)
- Carl von Ossietzky University Oldenburg (Project Partner)
- Logitech UK Ltd (Project Partner)
- Google Inc (Project Partner)
- RNID (Royal Natnl Inst for Deaf People) (Project Partner)
- Sonova AG (Project Partner)
- The British Broadcasting Corporation (BBC) (Project Partner)
Publications
Akeroyd M
(2024)
The Clarity & Cadenza Challenges
Akeroyd M
(2024)
Development of the 2nd Cadenza challenge for improving music listening for people with a hearing loss
in The Journal of the Acoustical Society of America
Akeroyd M.
(2023)
The Clarity & Cadenza Challenges
in Proceedings of Forum Acusticum
Bannister S
(2024)
Muddy, muddled, or muffled? Understanding the perception of audio quality in music by hearing aid users.
in Frontiers in psychology
Dabike G.R.
(2023)
The First Cadenza Signal Processing Challenge: Improving Music for Those With a Hearing Loss
in CEUR Workshop Proceedings
Firth J
(2023)
A systematic review of measurements of real-world interior car noise for the "Cadenza" machine-learning project
in The Journal of the Acoustical Society of America
Roa Dabike G
(2024)
The cadenza woodwind dataset: Synthesised quartets for music information retrieval and machine learning
in Data in Brief
Whitmer W
(2024)
Lyric intelligibility of musical segments for older individuals with hearing loss
in The Journal of the Acoustical Society of America
| Description | Organised the first open series of machine learning challenge for music processing and hearing loss. Challenges so far: CAD1, ICASSP24 and CAD2. This was part of developing a new research area and growing the number of researchers considering hearing loss and music technology. The results were presented at the first Cazenda Workshop in 2023 and ICASSP 2024. |
| Exploitation Route | The machine learning challenges have produced new data for the training of machine learning algorithms that include the effects of hearing loss on music. Cadenza has also published software tools to allow others to build systems. We have also worked with a sensory panel of people with hearing loss to define what the end-users want from the music processing. The challenges themselves are open benchmarks to promote further research. These all lower entry into the area of music, machine learning and hearing loss. |
| Sectors | Creative Economy Digital/Communication/Information Technologies (including Software) Healthcare |
| URL | https://cadenzachallenge.org/ |
| Description | We have run three competitions aimed at improving the signal processing in hearing aids and consumer devices for music. Any research group around the world can enter. This has created open-source databases, software tools and benchmark scenarios to promote further research in this area. In developing the challenges, we used a sensory panel of people with hearing loss to define what the target end-users want, to shape our work and future research by others. |
| First Year Of Impact | 2023 |
| Sector | Creative Economy,Digital/Communication/Information Technologies (including Software),Healthcare |
| Impact Types | Cultural Societal Economic |
| Title | Cadenza Challenge (CAD1): Submission audio samples for listening to music over the headphones challenge |
| Description | Cadenza This is the submission data for the listening over headphones task from the First Cadenza Machine Learning Challenge (CAD1). The Cadenza Challenges are improving music production and processing for people with a hearing loss. According to The World Health Organization, 430 million people worldwide have a disabling hearing loss. Studies show that not being able to understand lyrics is an important problem to tackle for those with hearing loss. Consequently, this task is about improving the intelligibility of lyrics when listening to pop/rock over headphones. But this needs to be done without losing too much audio quality - you can't improve intelligibility just by turning off the rest of the band! We will be using one metric for intelligibility and another metric for audio quality, and giving you different targets to explore the balance between these metrics. Please see the Cadenza website for a full description of the data |
| Type Of Material | Database/Collection of data |
| Year Produced | 2024 |
| Provided To Others? | Yes |
| Impact | Recently uploaded too soon for impact |
| URL | https://zenodo.org/doi/10.5281/zenodo.13271524 |
| Title | Cadenza Challenge (CAD1): databases for the First Cadenza Challenge - Task1 |
| Description | Cadenza This is the training, validation and evaluation data for the First Cadenza Challenge - Task 1. The Cadenza Challenges are improving music production and processing for people with a hearing loss. According to The World Health Organization, 430 million people worldwide have a disabling hearing loss. Studies show that not being able to understand lyrics is an important problem to tackle for those with hearing loss. Consequently, this task is about improving the intelligibility of lyrics when listening to pop/rock over headphones. But this needs to be done without losing too much audio quality - you can't improve intelligibility just by turning off the rest of the band! We will be using one metric for intelligibility and another metric for audio quality, and giving you different targets to explore the balance between these metrics. Please see the Cadenza website for a full description of the data |
| Type Of Material | Database/Collection of data |
| Year Produced | 2024 |
| Provided To Others? | Yes |
| Impact | Used in Cadenza open machine learning challenges |
| URL | https://zenodo.org/doi/10.5281/zenodo.13285383 |
| Title | Cadenza Challenge (CAD1): databases for the First Cadenza Challenge - Task2 |
| Description | This is the training, validation and evaluation data for the First Cadenza Challenge - Task 2 listening music in a car. The Cadenza Challenges are improving music production and processing for people with a hearing loss. According to The World Health Organization, 430 million people worldwide have a disabling hearing loss. Studies show that not being able to understand lyrics is an important problem to tackle for those with hearing loss. Consequently, this task is about improving the intelligibility of lyrics when listening to pop/rock over headphones. But this needs to be done without losing too much audio quality - you can't improve intelligibility just by turning off the rest of the band! We will be using one metric for intelligibility and another metric for audio quality, and giving you different targets to explore the balance between these metrics. Please see the Cadenza website for a full description of the data File description cadenza_cad1_task2_core.v1_1.tar.gz: Audio and metadata files for training and validation. cadenza_cad1_task2_evaluation.v1_1.tar.gz: Audio and metadata files for evaluation |
| Type Of Material | Database/Collection of data |
| Year Produced | 2024 |
| Provided To Others? | Yes |
| Impact | Too early for impact |
| URL | https://zenodo.org/doi/10.5281/zenodo.13329971 |
| Title | Cadenza Challenge (CAD2): databases for lyric intelligibility task |
| Description | Cadenza This is the training and validation data for the lyric intelligibility task from the Second Cadenza Machine Learning Challenge (CAD2). The Cadenza Challenges are improving music production and processing for people with a hearing loss. According to The World Health Organization, 430 million people worldwide have a disabling hearing loss. Studies show that not being able to understand lyrics is an important problem to tackle for those with hearing loss. Consequently, this task is about improving the intelligibility of lyrics when listening to pop/rock over headphones. But this needs to be done without losing too much audio quality - you can't improve intelligibility just by turning off the rest of the band! We will be using one metric for intelligibility and another metric for audio quality, and giving you different targets to explore the balance between these metrics. Please see the Cadenza website for a full description of the data |
| Type Of Material | Database/Collection of data |
| Year Produced | 2024 |
| Provided To Others? | Yes |
| Impact | Only recently published |
| URL | https://zenodo.org/doi/10.5281/zenodo.12685819 |
| Title | Cadenza Challenge (CAD2): databases for rebalancing classical music task |
| Description | Cadenza Please, cite CadenzaWoodwind as Gerardo Roa Dabike , Trevor J. Cox , Alex J. Miller , Bruno M. Fazenda , Simone Graetzer , Rebecca R. Vos , Michael A. Akeroyd , Jennifer Firth , William M. Whitmer , Scott Bannister , Alinka Greasley , Jon P. Barker , The Cadenza Woodwind Dataset: Synthesised Quartets for Music Information Retrieval and Machine Learning, Data in Brief (2024), doi: https://doi.org/10.1016/j.dib.2024.111199 This is the training and validation data for the rebalancing classic music task from the Second Cadenza Machine Learning Challenge (CAD2). The Cadenza Challenges are improving music production and processing for people with a hearing loss. According to The World Health Organization, 430 million people worldwide have a disabling hearing loss. Hearing aid users report several issues when listening to music, including distortion in the bass, difficulties in perceiving the full range of the music, especially high-frequency pitches, and a tendency to miss the impact of quieter parts of compositions [1]. In a pilot study, we found giving listeners sliders to allow them to rebalance different instruments in a classical music ensemble was desirable. Overview of files: CadenzaWoodwind. Synthesized dataset of small ensembles of woodwind instruments for training and validation. EnsembleSet_Mix_1. A subset of the synthesised EnsembleSet [7] for training and validation (Mix_1 render). Real Data for Tuning: Stereo_Reverb_Real_Data_For_Tuning.zip. metadata.zip contains audiograms, scene details, target gains and compressor settings. The audio files are in FLAC format in the .zip archives. The json files contain metadata. More details below. |
| Type Of Material | Database/Collection of data |
| Year Produced | 2024 |
| Provided To Others? | Yes |
| Impact | Only just uploaded |
| URL | https://zenodo.org/doi/10.5281/zenodo.12664931 |
| Title | Cadenza Challenge ICASSP 2024 (ICASSP24): Submission audio samples for the ICASSP 2024 Cadenza Grand Challenge - Baseline systems |
| Description | Cadenza This is the baseline submission data for the ICASSP 2024 Cadenza Grand Challenge (ICASSP24). The Cadenza Challenges are improving music production and processing for people with a hearing loss. According to The World Health Organization, 430 million people worldwide have a disabling hearing loss. Studies show that not being able to understand lyrics is an important problem to tackle for those with hearing loss. Consequently, this task is about improving the intelligibility of lyrics when listening to pop/rock over headphones. But this needs to be done without losing too much audio quality - you can't improve intelligibility just by turning off the rest of the band! We will be using one metric for intelligibility and another metric for audio quality, and giving you different targets to explore the balance between these metrics. Please see the Cadenza website for a full description of the data |
| Type Of Material | Database/Collection of data |
| Year Produced | 2024 |
| Provided To Others? | Yes |
| Impact | Too early for impact |
| URL | https://zenodo.org/doi/10.5281/zenodo.13285030 |
| Title | Muddy, Muddled, or Muffled? Understanding the Perception of Audio Quality in Music by Hearing Aid Users |
| Description | Underlying primary research data for a sensory evaluation study, focussing on the development of perceptual attributes for music audio quality from the perspective of hearing aid users. Research data include three .csv datasets corresponding to an individual elicitation task (single-word descriptors to describe music samples), background questionnaire (e.g., demographics, other characteristics), and a follow-up task (quantitative ratings relating to use of perceptual attributes of audio quality). These files are accompanied by a guidance document to navigate dataset variables. Finally, the research data include music samples (original and processed) utilised within the study. Part of the Cadenza ProjectPlease cite:Bannister S., Greasley A. E., Cox T. J., Akeroyd M. A., Barker J., Fazenda B., Firth J., Graetzer S. N., Roa Dabike G., Vos R. R., & Whitmer W. M. (2024). Muddy, muddled, or muffled? Understanding the perception of audio quality in music by hearing aid users. Frontiers in Psychology, 15: 1310176. Doi: 10.3389/fpsyg.2024.1310176.Music samples (original and processed) included in this research data are taken from the Medley DB database, with the following citations:Bittner, R., Salamon, J., Tierney, M., Mauch, M., Cannam, C., and Bello, J. P. (2014). MedleyDB: A Multitrack Dataset for Annotation-Intensive MIR Research. Taipei, Taiwan: 15th International Society for Music Information Retrieval Conference..Bittner, R., Wilkins, J., Yip, H., & Bello, J. (2016). MedleyDB 2.0: New Data and a System for Sustainable Data Collection. New York, NY, USA: International Conference on Music Information Retrieval (ISMIR-16).Medley DB materials are available under a CC-BY-NC-SA 4.0 International License. However, inclusion of original and processed music samples in this data release was also given consent by the authors of the database, on 2nd Monday October 2023. |
| Type Of Material | Database/Collection of data |
| Year Produced | 2024 |
| Provided To Others? | Yes |
| URL | https://salford.figshare.com/articles/dataset/Muddy_Muddled_or_Muffled_Understanding_the_Perception_... |
| Title | Muddy, Muddled, or Muffled? Understanding the Perception of Audio Quality in Music by Hearing Aid Users |
| Description | Underlying primary research data for a sensory evaluation study, focussing on the development of perceptual attributes for music audio quality from the perspective of hearing aid users. Research data include three .csv datasets corresponding to an individual elicitation task (single-word descriptors to describe music samples), background questionnaire (e.g., demographics, other characteristics), and a follow-up task (quantitative ratings relating to use of perceptual attributes of audio quality). These files are accompanied by a guidance document to navigate dataset variables. Finally, the research data include music samples (original and processed) utilised within the study. Part of the Cadenza ProjectPlease cite:Bannister S., Greasley A. E., Cox T. J., Akeroyd M. A., Barker J., Fazenda B., Firth J., Graetzer S. N., Roa Dabike G., Vos R. R., & Whitmer W. M. (2024). Muddy, muddled, or muffled? Understanding the perception of audio quality in music by hearing aid users. Frontiers in Psychology, 15: 1310176. Doi: 10.3389/fpsyg.2024.1310176.Music samples (original and processed) included in this research data are taken from the Medley DB database, with the following citations:Bittner, R., Salamon, J., Tierney, M., Mauch, M., Cannam, C., and Bello, J. P. (2014). MedleyDB: A Multitrack Dataset for Annotation-Intensive MIR Research. Taipei, Taiwan: 15th International Society for Music Information Retrieval Conference..Bittner, R., Wilkins, J., Yip, H., & Bello, J. (2016). MedleyDB 2.0: New Data and a System for Sustainable Data Collection. New York, NY, USA: International Conference on Music Information Retrieval (ISMIR-16).Medley DB materials are available under a CC-BY-NC-SA 4.0 International License. However, inclusion of original and processed music samples in this data release was also given consent by the authors of the database, on 2nd Monday October 2023. |
| Type Of Material | Database/Collection of data |
| Year Produced | 2024 |
| Provided To Others? | Yes |
| Impact | Data underpinning paper. |
| URL | https://salford.figshare.com/articles/dataset/Muddy_Muddled_or_Muffled_Understanding_the_Perception_... |
| Title | pyclarity |
| Description | pyclarity is a software suite for machine learning challenges to enhance hearing-aid signal processing and to better predict how people perceive speech-in-noise (Clarity) and speech-in-music (Cadenza). Files can be accessed via the Related Materials links below. |
| Type Of Technology | Software |
| Year Produced | 2023 |
| Impact | Used in open machine learning challenges to improve hearing aids for speech and music |
| URL | https://orda.shef.ac.uk/articles/software/pyclarity/23230694/1 |
| Description | Appearance on BBC Radio 4 Sliced Bread on Hearing Aids |
| Form Of Engagement Activity | A press release, press conference or response to a media enquiry/interview |
| Part Of Official Scheme? | No |
| Geographic Reach | International |
| Primary Audience | Media (as a channel to the public) |
| Results and Impact | Programme dedicated to hearing aids that discussed both Clarity and Cadenza project. |
| Year(s) Of Engagement Activity | 2025 |
| URL | https://www.bbc.co.uk/sounds/play/m00282qp |
| Description | Appearance on Start the Week |
| Form Of Engagement Activity | A press release, press conference or response to a media enquiry/interview |
| Part Of Official Scheme? | No |
| Geographic Reach | National |
| Primary Audience | Media (as a channel to the public) |
| Results and Impact | Appearance on BBC Radio 4 Start the Week about Acoustics included a brief discussion of Cadenza, Music and Hearing Aids. |
| Year(s) Of Engagement Activity | 2024 |
| Description | Patient group workshop |
| Form Of Engagement Activity | Participation in an activity, workshop or similar |
| Part Of Official Scheme? | No |
| Geographic Reach | National |
| Primary Audience | Study participants or study members |
| Results and Impact | Ran two sensory panels of 12 people with hearing loss as upstream engagement to discover the issues important to them and shape the research. |
| Year(s) Of Engagement Activity | 2022,2023 |
| URL | http://cadenzachallenge.org |
| Description | The 1st Cadenza Workshop on Machine Learning Challenges to Improve Music for People with Hearing Loss (Cadenza-2023) |
| Form Of Engagement Activity | Participation in an activity, workshop or similar |
| Part Of Official Scheme? | No |
| Geographic Reach | International |
| Primary Audience | Other audiences |
| Results and Impact | The Cadenza workshops are designed to stimulate the development of systems to improve music for people with a hearing loss. The aim of this virtual workshop was to report on the First Cadenza Challenge, the first ever machine learning challenge targeted to improve music for those with a hearing impairment. |
| Year(s) Of Engagement Activity | 2023 |
| URL | https://cadenzachallenge.org/cadenza2023-workshop/ |
| Description | Workshop at the 4th Virtual Conference on Computational Audiology (VCCA2023) |
| Form Of Engagement Activity | Participation in an activity, workshop or similar |
| Part Of Official Scheme? | No |
| Geographic Reach | International |
| Primary Audience | Other audiences |
| Results and Impact | Demos and discussion around how can we process and remix music, so it sounds better for those with a hearing loss. This upstream engagement shaped the machine learning challenges in the Cadenza project (EPSRC project EnhanceMusic). Getting input from hearing aid experts, audiologists and academics. |
| Year(s) Of Engagement Activity | 2023 |
| URL | https://computationalaudiology.com/events/vcca2023/ |
