CausalXRL: Causal eXplanations in Reinforcement Learning

Lead Research Organisation: University of Sheffield
Department Name: Computer Science

Abstract

Deep reinforcement learning (RL) systems are approaching or surpassing human-level performance in specific domains, from games to decision support to continuous control, albeit in non-critical environments, and usually learning via random explorations. Despite these prodigious achievements, many applications cannot be considered today because we need to understand and explain how these AI systems make their decisions before letting them interact with, and possibly impact on, human beings and society.

There are two main obstacles for AI agents to explain their decisions: they have to be able to provide it at a level human beings can understand, and they have to deal with causal relations rather than statistical correlations. Hence, we believe the key to explainable AI, in particular for decision support, is to build or learn causal models of the system being intervened upon. Thus, instead of standard machine learning and reinforcement learning networks, we will leverage the new science of causal inference to equip deep RL systems with the ability to learn, plan with, and justifiably explore cause-effect relationships in their environment. RL systems based on this novel CausalXRL architecture will provide cause-effect and counterfactual justifications for their suggested actions, allowing them to fulfill the right to an explanation in human-centric environments.

We will implement the CausalXRL architecture as a bio-plausible (neuromorphic) algorithm to enable its deployment in resource-limited, e.g., mobile, environments. We will demonstrate the broad applicability and impact of CausalXRL on several use cases, ranging from neuro-rehabilitation to intensive care, farming and education.

Publications

10 25 50
 
Description Our CHIST-ERA partners 
Organisation Inria research centre Lille - Nord Europe
Country France 
Sector Public 
PI Contribution Inria Lille and the University of Vienna are our CHIST-ERA partners in France and Austria correspondingly, linked to this award. Sheffield collaborates with NWO to provide expertise on (more) biologically plausible versions of a machine learning technique named reinforcement learning. We work in close collaboration with the partners to develop a practical demonstrator. The demonstrator combines model-based reinforcement learning and variational autoencoders. The latter is an unsupervised technique that allows for dimensionality reduction by discovering latent representations. This combination enables reinforcement learning methods to learn more efficiently.
Collaborator Contribution The French partner provides considerable expertise on sequential decision making under uncertainty, including classical reinforcement learning. The specific version of autoencoders is guided by our third partner in Austria, an expert on causality, as we are looking for causal representations. The Austrian partner also works independently to provide an appropriate causal paradigm on which the French partner will develop novel reinforcement learning versions.
Impact There are no outcomes yet. The kickoff meeting for the project was on December 2021. The partners offer different expertise but fall largely in the domain of computer science.
Start Year 2021
 
Description Our CHIST-ERA partners 
Organisation Netherlands Organisation for Scientific Research (NWO)
Department National Research Institute for Mathematics and Computer Science
Country Netherlands 
Sector Academic/University 
PI Contribution Inria Lille and the University of Vienna are our CHIST-ERA partners in France and Austria correspondingly, linked to this award. Sheffield collaborates with NWO to provide expertise on (more) biologically plausible versions of a machine learning technique named reinforcement learning. We work in close collaboration with the partners to develop a practical demonstrator. The demonstrator combines model-based reinforcement learning and variational autoencoders. The latter is an unsupervised technique that allows for dimensionality reduction by discovering latent representations. This combination enables reinforcement learning methods to learn more efficiently.
Collaborator Contribution The French partner provides considerable expertise on sequential decision making under uncertainty, including classical reinforcement learning. The specific version of autoencoders is guided by our third partner in Austria, an expert on causality, as we are looking for causal representations. The Austrian partner also works independently to provide an appropriate causal paradigm on which the French partner will develop novel reinforcement learning versions.
Impact There are no outcomes yet. The kickoff meeting for the project was on December 2021. The partners offer different expertise but fall largely in the domain of computer science.
Start Year 2021
 
Description Our CHIST-ERA partners 
Organisation University of Vienna
Country Austria 
Sector Academic/University 
PI Contribution Inria Lille and the University of Vienna are our CHIST-ERA partners in France and Austria correspondingly, linked to this award. Sheffield collaborates with NWO to provide expertise on (more) biologically plausible versions of a machine learning technique named reinforcement learning. We work in close collaboration with the partners to develop a practical demonstrator. The demonstrator combines model-based reinforcement learning and variational autoencoders. The latter is an unsupervised technique that allows for dimensionality reduction by discovering latent representations. This combination enables reinforcement learning methods to learn more efficiently.
Collaborator Contribution The French partner provides considerable expertise on sequential decision making under uncertainty, including classical reinforcement learning. The specific version of autoencoders is guided by our third partner in Austria, an expert on causality, as we are looking for causal representations. The Austrian partner also works independently to provide an appropriate causal paradigm on which the French partner will develop novel reinforcement learning versions.
Impact There are no outcomes yet. The kickoff meeting for the project was on December 2021. The partners offer different expertise but fall largely in the domain of computer science.
Start Year 2021