Environment Representation is one major category in our taxonomy. The core research content of environment representation is to capture the variations of environments from the distribution of interests. Such variations can be the inner factors that determine the dynamics, reward functions and other aspects of MDP.
Since these factors are usually inaccessable directly, to achieve the purpose of generalizing well and staying robust in face of such changing variations, environment representation is learned to infer the underlying factors from accessible data (e.g., interaction experiences).
This repo contains representative research works of TJU-RL-Lab on the topic of RL with Environment Representation.
As mentioned above, we aim at learning good representations of underlying variations of environments with accessible data. Obviously, a reasonable type of such data is transition experiences {s,a,r,s'}, also called context, collected during the interaction of RL agent in the environment, since interaction experiences can implicitly reflect the information of environment characteristics, i.e., specific underlying variations.
We provide a conceptual illustration of Context-based Environment Representation and Generalization below:
The representations of environment are learned by encoding context data and extracting information according to specific context learning objectives. Afterwards, Context-conditioned RL functions, e.g., 𝑄(𝑠,𝑎,𝑧), 𝜋(⋅|𝑠,𝑧) are trained, to make the RL agent aware of the environment representation (i.e., variations ideally). Therefore, with such context-conditioned RL functions, implicit generalization or explicit adaptation among environments can be carried out.
This repo will be constantly updated to include new researches made by TJU-RL-Lab. (The development of this repo is in progress at present.)
Method | Is Contained | Author | Publication | Link |
---|---|---|---|---|
CCM | ✅ | Haotian Fu | AAAI 2021 | https://ojs.aaai.org/index.php/AAAI/article/view/16914 |
PAnDR | ✅ | Tong Sang | IJCAI 2022 | https://arxiv.org/abs/2204.02877 |
The algorithms in this repo are all implemented python 3.5 (and versions above). Tensorflow 1.x and PyTorch are the main DL code frameworks we adopt in this repo with different choices in different algorithms.
Note that the algorithms contained in this repo may not use all the same environments. Please check the README of specific algorithms for detailed installation guidance.
Here we provide a useful list of representative related works on environment (or task) representation in RL.
- Kate Rakelly, Aurick Zhou, Chelsea Finn, Sergey Levine, Deirdre Quillen. Efficient Off-Policy Meta-Reinforcement Learning via Probabilistic Context Variables. ICML 2019
- Haotian Fu, Hongyao Tang, Jianye Hao, Chen Chen, Xidong Feng, Dong Li, Wulong Liu. Towards Effective Context for Meta-Reinforcement Learning: an Approach based on Contrastive Learning. AAAI 2021
- Kimin Lee, Younggyo Seo, Seunghyun Lee, Honglak Lee, Jinwoo Shin. Context-aware Dynamics Model for Generalization in Model-Based Reinforcement Learning. ICML 2020
- Bernie Wang, Simon Xu, Kurt Keutzer, Yang Gao, Bichen Wu. Improving Context-Based Meta-Reinforcement Learning with Self-Supervised Trajectory Contrastive Learning. arXiv:2103.06386
- Wenxuan Zhou, Lerrel Pinto, Abhinav Gupta. Environment Probing Interaction Policies. ICLR (Poster) 2019
- Roberta Raileanu, Max Goldstein, Arthur Szlam, Rob Fergus. Fast Adaptation via Policy-Dynamics Value Functions. ICML 2020
- Yujing Hu, Weixun Wang, Hangtian Jia, Yixiang Wang, Yingfeng Chen, Jianye Hao, Feng Wu, Changjie Fan. Learning to Utilize Shaping Rewards: A New Approach of Reward Shaping. NIPS 2020
- Daniel S. Brown, Wonjoon Goo, Prabhat Nagarajan, Scott Niekum. Extrapolating Beyond Suboptimal Demonstrations via Inverse Reinforcement Learning from Observations. ICML 2019
- Alexander C. Li, Lerrel Pinto, Pieter Abbeel. Generalized Hindsight for Reinforcement Learning. NIPS 2020