Conversational AI Audiology: Remote, natural and automated testing of hearing and fitting
Lead Research Organisation:
The University of Manchester
Department Name: School of Health Sciences
Abstract
Conversational artificial intelligence (AI), i.e. technology that a patient can talk to, has the potential to remove barriers from access to healthcare and in particular hearing healthcare. Imagine that people who worry about their hearing would not need to visit a GP or an audiologist for a first diagnosis but could just ask their home assistant "Please check my hearing". This could lead to an increasing awareness of hearing loss and uptake of hearing aids by the 1.5 billion people (2.5 billion in 2050) who suffer from a hearing loss. It is estimated that at least 6 million people in the UK would benefit from a hearing aid but only 2 million have one, and uptake among them is low and slow. A hearing test that takes place as a conversation with a technical device has further advantages to conventional technologies: First, it tests real-world sounds (speech) rather than the audibility of faint synthetic sounds (pure tones in an audiogram). Second, a conversational AI system can directly simulate the speech that would be delivered through a hearing aid and thus quantify the benefit that the best hearing-aid setting would give. Third, remote testing does not only remove barriers in accessing healthcare but is also of benefit to vulnerable patients. In this project we will apply text-to-speech (TTS) and automatic speech recognition (ASR) to hearing tests in order to achieve these goals. TTS and ASR will allow the patient to interact easily with the test system, which is an additional benefit for those who struggle with graphical user interfaces, in particular some elderly patients. The communication between patient and AI will be driven by the goals to have a more natural interaction than in conventional tests and to characterise a hearing loss in short time. The candidate should have an interest in healthcare and is expected to have a background in deep learning and other fields of artificial intelligence, obtained from a first degree in engineering, computer science, mathematics, physics or similar. The candidate will acquire skills in TTS, ASR and active learning. Collaborations with other young researchers and activities like journal clubs will contribute to establishing a basic knowledge in audiology and clinical testing. The candidate will produce research outputs that are published in machine-learning conferences (NeurIPS, ICML, Interspeech, ICASSP) and leading hearing journals (Trends in Hearing, IJA, Hearing Research). Further references: Schlittenlacher, J., Baer, T. (2021) Text-to-speech for the hearing impaired, arXiv, 2012.02174
People |
ORCID iD |
Kevin Munro (Primary Supervisor) | |
Mohsen Fatehifar (Student) |
Studentship Projects
Project Reference | Relationship | Related To | Start | End | Student Name |
---|---|---|---|---|---|
MR/W007428/1 | 30/09/2022 | 29/09/2028 | |||
2776430 | Studentship | MR/W007428/1 | 30/09/2022 | 29/09/2026 | Mohsen Fatehifar |