Inverse Reinforcement Learning

Inverse Reinforcement Learning (IRL) is the machine learning paradigm concerned with inferring the latent reward function of an agent based on its observed behavior. Formally, given a Markov Decision Process (MDP) without a specified reward signal and a set of expert demonstrations (state-action trajectories), IRL seeks to recover the underlying utility function that the expert is assumed to be optimally maximizing. This effectively inverts the standard reinforcement learning problem: rather than deriving a policy from a known reward, it derives the reward structure that best explains the observed policy.

TitleAuthor / YearThemeComment
Maximum Entropy IRLZiebart et al. (2008)AlgorithmMethodological Paper
Deep Maximum Entropy IRLWulfmeier et al. (2015)AlgorithmMethodological Paper
Adversarial Inverse Reinforcement LearningFu et al. (2018)AlgorithmMethodological Paper
Inverse soft-Q Learning for Imitation (Environment Free)Garg et al. (2022)AlgorithmMethodological Paper
Variational IRL (Environment Free)Qureshi et al. (2019)AlgorithmMethodological Paper
Multi-Agent Adversarial IRLYu et al. (2019)Algorithmic EnhancementMethodological Paper
Context-aware IRLLiu et al. (2025)Modeling human behavior using IRLApplication Paper
IRL for modeling reservoir operationsGiuliani and Castelletti (2024)Modeling human behavior using IRLApplication Paper
Multiple Expert and Non-stationarity in IRLLikmeta et al. (2021)Modeling human behavior using IRLApplication Paper
Advances and Applications in IRLDeshpande et al. (2025)Algorithms and ApplicationLiterature Review

© 2024. All rights reserved.

Powered by Hydejack v9.2.1