Wanqianxn / usfa
Implementation of USFAs: https://arxiv.org/pdf/1812.07626.pdf
☆9Updated 6 years ago
Alternatives and similar repositories for usfa:
Users that are interested in usfa are comparing it to the libraries listed below
- ☆31Updated 5 years ago
- Project on Successor Features in Deep Reinforcement Learning and Transfer Learning☆24Updated 7 years ago
- ☆53Updated last year
- Simple maze environments using mujoco-py☆54Updated last year
- ☆28Updated 3 years ago
- Implementation of the Option-Critic Architecture☆39Updated 6 years ago
- Learning Laplacian Representations in Reinforcement Learning☆17Updated 4 years ago
- Pessimistic Bootstrapping for Uncertainty-Driven Offline Reinforcement Learning☆28Updated 3 years ago
- Random parameter environments using gym 0.7.4 and mujoco-py 0.5.7☆20Updated 6 years ago
- Maximum Entropy Population Based Training for Zero-Shot Human-AI Coordination☆28Updated 2 years ago
- ☆29Updated 2 years ago
- Submission for MAVEN: Multi-Agent Variational Exploration☆57Updated 3 years ago
- Proximal Policy Option-Critic☆22Updated 6 years ago
- ☆26Updated 2 years ago
- Pytorch code for "Learning Belief Representations for Imitation Learning in POMDPs" (UAI 2019)☆18Updated 2 years ago
- The Implementation of "Machine Theory of Mind", ICML 2018☆24Updated 3 years ago
- Code for FOCAL Paper Published at ICLR 2021☆52Updated last year
- ☆55Updated 2 years ago
- ☆42Updated 3 years ago
- OPE Tools based on Empirical Study of Off Policy Policy Estimation paper.☆61Updated 2 years ago
- A reusable framework for successor features for transfer in deep reinforcement learning using keras.☆43Updated 3 years ago
- Learning bisimulation metrics for control, particularly suited to sparse reward settings☆10Updated 2 years ago
- ☆60Updated 6 years ago
- Code accompanying NeurIPS 2019 paper: "Distributional Policy Optimization - An Alternative Approach for Continuous Control"☆22Updated 5 years ago
- ☆53Updated 4 years ago
- COOM: Benchmarking Continual Reinforcement Learning on Doom☆17Updated last month
- Offline Risk-Averse Actor-Critic (O-RAAC). A model-free RL algorithm for risk-averse RL in a fully offline setting☆35Updated 4 years ago
- ☆31Updated 4 years ago
- Safe Option-Critic: Learning Safety in the Option-Critic Architecture☆20Updated 6 years ago
- Inverse Reinforcement Learning via State Marginal Matching, CoRL 2020☆45Updated last year