šŸ“£ Help Shape the Future of UKRI's Gateway to Research (GtR)

We're improving UKRI's Gateway to Research and are seeking your input! If you would be interested in being interviewed about the improvements we're making and to have your say about how we can make GtR more user-friendly, impactful, and effective for the Research and Innovation community, please email gateway@ukri.org.

GraphNEx: Graph Neural Networks for Explainable Artificial Intelligence

Lead Research Organisation: Queen Mary University of London
Department Name: Sch of Electronic Eng & Computer Science

Abstract

GraphNEx will contribute a graph-based framework for developing inherently explainable AI. Unlike current AI systems that utilise complex networks to learn high-dimensional, abstract representations of data, GraphNEx embeds symbolic meaning within AI frameworks. We will combine semantic reasoning over knowledge bases with simple modular learning on new data observations, to adaptively evolve the graphical knowledge base. The core concept is to decompose the monolithic block of highly connected layers of deep learning architectures into many smaller, simpler units. Units perform a precise task with an interpretable output, associating a value or vector to nodes. Nodes will be connected depending on their (learned) similarity, correlation in the data or closeness in some space. We will employ the concepts and tools from graph signal processing and graph machine learning (e.g. Graph Neural Networks) to extrapolate semantic concepts and meaningful relationships from sub-graphs (concepts) within the knowledge base that can be used for semantic reasoning. By enforcing sparsity and domain-specific priors between concepts we will promote human interpretability. The integration of game-based user feedback will ensure that explanations (and therefore core mechanisms of the AI system) are understandable and relevant to humans. We will validate the GraphNEx framework, including both model performance and the performance of the explanations, with two application scenarios in which transparency and trust are critical: System Genetics to assist the discovery of clinically and biologically relevant concepts on large, multimodal genetic and genomic datasets; and Privacy Protection to safeguard personal information from unwanted non-essential inferences from multimedia data (images, video and audio) as well as to support informed consent. These applications, for which understanding the decision process is paramount for acquisition of new knowledge and for trust by the users, will demonstrate the utility of the GraphNEx framework to adapt to contexts where data are heterogeneous, incomplete, noisy and time-varying, providing new inherently explainable AI models, and a means to explain existing AI systems via post-hoc analyses.
 
Description Identification of key features that distiguish a private image from a public one, use of graph processing for image and audio data, identification of human-interpretable features for the task at hand.
Exploitation Route Further use to design feature representations for addressing digital privacy protection. Further use of the graph architecture for classification problems
Sectors Digital/Communication/Information Technologies (including Software)

Education

 
Title Explaining models relating objects and privacy 
Description To explain the decision of privacy classification models, we use feature-attribution to identify and quantify which objects (and which of their features) are more relevant to privacy classification with respect to a reference input (i.e., no objects localised in an image) predicted as public. We show that the presence of the person category and its cardinality is the main factor for the privacy decision. Therefore, these models mostly fail to identify private images depicting documents with sensitive data, vehicle ownership, and internet activity, or public images with people (e.g., an outdoor concert or people walking in a public space next to a famous landmark). As baselines for future benchmarks, we also devise two strategies that are based on the person presence and cardinality and achieve comparable classification performance of the privacy models. 
Type Of Material Computer model/algorithm 
Year Produced 2024 
Provided To Others? Yes  
Impact Not yet available. 
URL https://github.com/graphnex/ig-privacy
 
Description ENSL-graphnex 
Organisation Ɖcole normale supĆ©rieure de Lyon (ENS Lyon)
Country France 
Sector Academic/University 
PI Contribution We are collaborating to develop techniques using machine learning on graphs to define a framework for discovery of rules from inputs to outputs to be used on the two case studies of the project Graphnex. This collaboration started 6 months ago and is expected to lead to results next year.
Collaborator Contribution ENSL is providing expertise in Graph Neural Networks (problem modelling and framework).
Impact N/A This collaboration started 6 months ago and is expected to lead to results next year.
Start Year 2021
 
Description Demo at the V&A museum 
Form Of Engagement Activity Participation in an activity, workshop or similar
Part Of Official Scheme? No
Geographic Reach International
Primary Audience Other audiences
Results and Impact Interactive demo of the Privacy use case at the Digital Design Weekend 2022 at the Victoria and Albert (V&A) museum in London. See https://www.youtube.com/watch?v=LR9Q_KTAhOs
Year(s) Of Engagement Activity 2022
URL https://www.youtube.com/watch?v=LR9Q_KTAhOs