clvoloshin / COBS
OPE Tools based on Empirical Study of Off Policy Policy Estimation paper.
☆61Updated 2 years ago
Alternatives and similar repositories for COBS:
Users that are interested in COBS are comparing it to the libraries listed below
- ☆85Updated 7 months ago
- ☆42Updated 3 years ago
- ☆60Updated 6 years ago
- Random parameter environments using gym 0.7.4 and mujoco-py 0.5.7☆20Updated 6 years ago
- Model-Based Offline Reinforcement Learning☆48Updated 4 years ago
- ☆53Updated last year
- Code for MOPO: Model-based Offline Policy Optimization☆173Updated 2 years ago
- ☆26Updated 5 years ago
- ☆42Updated 6 years ago
- ☆26Updated last year
- Code for demonstration example-task in RUDDER blog☆23Updated 4 years ago
- Safe Policy Improvement with Baseline Bootstrapping☆26Updated 4 years ago
- Implementation of "Sample-Efficient Deep Reinforcement Learning via Episodic Backward Update", NeurIPS 2019.☆17Updated 5 years ago
- Pessimistic Bootstrapping for Uncertainty-Driven Offline Reinforcement Learning☆28Updated 3 years ago
- on-policy optimization baselines for deep reinforcement learning☆29Updated 4 years ago
- ☆66Updated 4 years ago
- Official implementation of Causal Curiosity: RL Agents Discovering Self-supervised Experiments for Causal Representation Learning at ICML…☆38Updated 3 years ago
- ☆29Updated 2 years ago
- ☆91Updated last year
- ☆31Updated 5 years ago
- Official PyTorch implementation of "Uncertainty-Based Offline Reinforcement Learning with Diversified Q-Ensemble" (NeurIPS'21)☆75Updated 2 years ago
- Code for paper Causal Confusion in Imitation Learning☆45Updated 5 years ago
- Source for the sample efficient tabular RL submission to the 2019 NIPS workshop on Biological and Artificial RL☆23Updated 2 years ago
- This code implements Prioritized Level Replay, a method for sampling training levels for reinforcement learning agents that exploits the …☆84Updated 3 years ago
- Code for Diagnosing Bottlenecks in Deep Q-learning. Contains implementations of tabular environments plus solvers.☆19Updated 5 years ago
- Offline Risk-Averse Actor-Critic (O-RAAC). A model-free RL algorithm for risk-averse RL in a fully offline setting☆34Updated 4 years ago
- Invariant Causal Prediction for Block MDPs☆44Updated 4 years ago
- ☆28Updated 3 years ago
- ☆15Updated 4 years ago
- ☆193Updated last year