Causal Foundations of Reinforcement Learning
Lead Research Organisation:
University College London
Department Name: Computer Science
Abstract
Reinforcement learning is essentially a causal inference task, although much of traditional RL is done with fully controllable agents in which core causal techniques can only play a trivial role and hence have been mostly ignored. With the recognition that many multistage decision processes are hard to learn without a large number of trials, there is an increased need to use data collected under observational or known off-policy regimes. This project will focus on learning techniques to make the most of tapping on uncontrolled logged trajectories, assessing the fundamental questions of uncertainty and extrapolation problems when data is not generated by direct exploration of the environment by the learner, and how sensitivity analysis techniques can highlight the brittleness of our knowledge in regions of the trajectory space where new interventions need to be done. If AI research wants to engage with real-world problems of sequential planning, it needs to go deeper on methodology that addresses how to combine multiple data sources and structural assumptions, as data-intensive reinforcement learning is not feasible in many problems.
People |
ORCID iD |
Ricardo Bezerra Silva (Primary Supervisor) | |
Jakob Zeitler (Student) |
Studentship Projects
Project Reference | Relationship | Related To | Start | End | Student Name |
---|---|---|---|---|---|
EP/S021566/1 | 31/03/2019 | 29/09/2027 | |||
2252836 | Studentship | EP/S021566/1 | 22/09/2019 | 19/01/2024 | Jakob Zeitler |