Sequential decision making and credit assignment under uncertainty and partial observability is central to developing Intelligent Systems. Reinforcement Learning (RL) provides a general and powerful computational framework for sequential decision making. It involves an agent interacting with the environment to maximize a reward function by selecting actions.
Our research at the Institute of Machine Learning focuses on developing new algorithms and theory required to improve the state of the art in Reinforcement Learning. Credit assignment under delayed reward has been central to our work in recent years. We also actively pursue developing new function approximation methods for scaling Reinforcement Learning to high dimensional problems. Learning to take decisions based on stored data is another area of interest. We actively apply Reinforcement Learning to various applications including robotics, logistics, natural language processing and others.
recent publications in Reinforcement Learning:
- RRLLearning to Modulate pre-trained Models in RL2023
- Toward Semantic History Compression for Reinforcement Learning2022
- DeepRLInfODist: Online distillation with Informative rewards improves generalization in Curriculum Learning2022
- FMDMFoundation Models for History Compression in Reinforcement Learning2022
- CoLLAsA Dataset Perspective on Offline Reinforcement Learning2022
- CoLLAsReactive Exploration to Cope with Non-Stationarity in Lifelong Reinforcement Learning2022
- ICMLHistory Compression via Language Models in Reinforcement LearningIn 2022
- ICMLAlign-RUDDER: Learning From Few Demonstrations by Reward RedistributionarXiv preprint arXiv:2009.14108 2022
- arXivUnderstanding the Effects of Dataset Characteristics on Offline Reinforcement Learning2021
- Modern Hopfield Networks for Return Decomposition for Delayed RewardsIn Deep RL Workshop NeurIPS 2021 2021
- arXivConvergence Proof for Actor-Critic Methods Applied to PPO and RUDDER2020
- NeurIPSRUDDER: Return Decomposition for Delayed RewardsIn Advances in Neural Information Processing Systems 2019