📣 Help Shape the Future of UKRI's Gateway to Research (GtR)

We're improving UKRI's Gateway to Research and are seeking your input! If you would be interested in being interviewed about the improvements we're making and to have your say about how we can make GtR more user-friendly, impactful, and effective for the Research and Innovation community, please email gateway@ukri.org.

Hierarchical Deep Representations of Anatomy (HiDRA)

Lead Research Organisation: University of Lincoln
Department Name: School of Computer Science

Abstract

Understanding the functions of genes in animal models such as mice allows researchers to learn about their roles in human disease. High-throughput phenotyping is used to conduct broad assessments of gene function through a combination of qualitative and quantitative assays, which seek to measure or visualise specific anatomical structures or organ systems. It is essential to understanding genotype-phenotype relationships and has guided the development of therapeutic targets for developmental, cardiovascular, neurodegenerative, and sensory disorders. Skeletal phenotyping is particularly crucial from a public health standpoint, with musculoskeletal disorders being responsible for 12% of all general practitioner visits in the UK at a cost of 10.8 million working days and some £4.7 billion to the NHS per year.

Broad assessments of the skeleton are particularly laborious and subjective due to its anatomical complexity and the range of potential anomalies one might observe. In mouse phenotyping, plain x-ray images are routinely acquired from multiple viewpoints, orientations and scales to ensure complete coverage of the whole animal. Phenotypes are identified through manual inspection by domain experts, which is prohibitively time-consuming to perform at scale. The International Phenotyping Consortium (IMPC) comprises eight institutions that collect x-ray images of mice and annotate up 52 different phenotypes affecting skull, teeth, ribs, spine, pelvis and limbs. This represents a monumental task, generating some 166,000 annotated images from 34,000 animals to date. However, this body of data only represents 7,500 of the 20,000 genes phenotyped so far by the IMPC. The bottleneck of manual annotation is a daunting prospect for the project and represents an unmet need for automated image analysis methods within the life sciences community.

In recent years, convolutional neural networks (CNNs) have risen to prominence for their seemingly universal applicability to a wide range of image classification problems. This partly due to the fact they require little prior domain knowledge to implement, and the data require minimal pre-processing order to achieve state-of-the-art performance. However, there are nevertheless challenges that preclude their adoption for large-scale phenotyping. Traditional CNNs are not naturally suited to datasets where the number of input images is variable; an individual animal may be captured from one and up six different viewpoints in practice. Furthermore, CNNs trained to perform multiple tasks at once (i.e., one animal may exhibit multiple phenotypes) have little appreciation or knowledge of the relationships between the tasks due to anatomical proximity. Beyond the immediate use-case of CNNs for automation, there is an opportunity to leverage the internal representations learned by CNNs to perform large-scale data mining and support biological discovery.

HiDRA (Hierarchical Deep Representations of Anatomy) will address these challenges by developing a "multi-view-multi-task" approach that is robust to variations in the input data, shares information between anatomically-related tasks, and learns anatomy-specific feature representations for individual animals. Information fusion from multiple viewpoints will allow for any number of images to be provided and help to indicate which view was most informative for annotation. To account for the relative scarcity of individual phenotypes, a hierarchical training scheme will be developed to share information across related tasks according to anatomical proximity. The learned representations will also be subject to constraints that minimise correlations between anatomical structures, allowing for comparisons to be made between animals in an anatomy-specific fashion using data mining techniques. Among the outputs of this research will include computational tools made available to the wider life sciences community for analysis of x-ray data at any scale.

Publications

10 25 50
 
Description The aim of this research was to explore whether a single AI model can accurately detect a wide range of skeletal abnormalities from x-ray images of mice. Our research has shown that a hierarchical training regime improves the performance of the AI, as compared to direct training. By first training the model to characterise high-level abnormalities (e.g., normal vs abnormal), it can more reliably detect more fine-grained, anatomy-specific abnormalities (e.g., abnormal spine curvature).
Exploitation Route This general approach has broad applicability for high-throughput phenotyping, where there is likely to be a wide range of abnormalities that can be grouped under one or more common themes (e.g., anatomy, pathology).
Sectors Agriculture

Food and Drink

Digital/Communication/Information Technologies (including Software)

Healthcare

Pharmaceuticals and Medical Biotechnology

 
Title Hierarchical training for multi-label anatomical phenotyping 
Description The proposed methodology is designed to allow for complex annotations of anatomical phenotypes, whereby model training is carried out in a sequential, hierarchical fashion. Under this scheme, common anatomical structures are grouped and organised by increasing level of granularity. At level 1, an individual subject (mouse) is regarded as either "normal" or "abnormal". At level 2, the same binary (normal/abnormal) classification is applied to non-mutually exclusive categories such as "spine", "skull" and "limbs". At level 3 and beyond, these categories are broken down further "e.g., lumbar vertebrae". Our research has shown that this hierarchical training scheme yields superior performance to training at level 3 directly. 
Type Of Material Computer model/algorithm 
Year Produced 2024 
Provided To Others? No  
Impact The work was presented at MIUA 2024 and is currently being expanded into a full paper submission. 
 
Title Multiview image-based hierarchical classification 
Description The research employs Multiview image data for comprehensive representation of mice specimen. The dorsoventral and lateral view images were integrated for compact representation and hierarchical classification model was developed. The hierarchical classification was performed using ConvNext and Convolutional Autoencoder (CAE) backbones. The results demonstrated the efficacy of multi-view representation over any single view for hierarchical classification. 
Type Of Material Computer model/algorithm 
Year Produced 2025 
Provided To Others? No  
Impact A simple version was presented in MIUA2024 whereas a complete paper is ready for submission to journal. Moreover, we also aim to submit conference paper in MIUA 2025. 
 
Description Lincolnshire Heart Centre 
Organisation United Lincolnshire Hospitals NHS Trust
Country United Kingdom 
Sector Public 
PI Contribution Moazzam Jawaid (the PDRA associated with this award) has expertise in the quantification of coronary heart disease from CTCA images. Alongside the primary focus of this project, he has continued to pursue his interests in CTCA analysis (utilising skills learned under this award) which led to a publication in 2024. Since then, we have explored this new direction further and established a collaboration with researchers at Lincolnshire Heart Centre. It is our intention to submit one or more bids to support future research into automated plaque characterisation, as well as an EPSRC Doctoral Landscape Award studentship proposal.
Collaborator Contribution Our partners have provided the clinical motivation and highlighted an unmet need for technology capable of characterising radiographic plaque characteristics. An in-kind contribution in the form of clinical imaging data is expected, subject to DPIA and ethical approval.
Impact This interdisciplinary collaboration (computer science, interventional cardiology) is new, and has not yet yielded any tangible outcomes.
Start Year 2025