Preference Transformer: Modeling Human Preferences using Transformers for RL (ICLR2023 Accepted)
☆167Oct 15, 2023Updated 2 years ago
Alternatives and similar repositories for PreferenceTransformer
Users that are interested in PreferenceTransformer are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- Official codebase for "B-Pref: Benchmarking Preference-BasedReinforcement Learning" contains scripts to reproduce experiments.☆134Nov 3, 2021Updated 4 years ago
- PyTorch code accompanying the paper "Imitating Graph-Based Planning with Goal-Conditioned Policies" (ICLR 2023).☆21Mar 4, 2023Updated 3 years ago
- Listwise Reward Estimation for Offline Preference-based Reinforcement Learning (ICML 2024)☆17Jun 18, 2024Updated last year
- ☆43May 25, 2023Updated 2 years ago
- Jaehyung Kim et al's ACL 2023 paper on "infoVerse: A Universal Framework for Dataset Characterization with Multidimensional Meta-informat…☆16Jun 28, 2023Updated 2 years ago
- Meta-Learning with Self-Improving Momentum Target (NeurIPS 2022)☆23Oct 12, 2022Updated 3 years ago
- ☆37Apr 27, 2023Updated 2 years ago
- Guide Your Agent with Adaptive Multimodal Rewards (NeurIPS 2023 Accepted)☆33Sep 25, 2023Updated 2 years ago
- Learning Large-scale Neural Fields via Context Pruned Meta-Learning (NeurIPS 2023)☆28Sep 24, 2023Updated 2 years ago
- PyTorch implementations for Offline Preference-Based RL (PbRL) algorithms☆21Mar 24, 2025Updated 11 months ago
- ☆10Mar 11, 2024Updated 2 years ago
- Official Codebase for TMLR 2023, Benchmarks and Algorithms for Offline Preference-Based Reward Learning☆20Dec 30, 2022Updated 3 years ago
- Subtask-Aware Visual Reward Learning from Segmented Demonstrations (ICLR 2025 accepted)☆18Apr 11, 2025Updated 11 months ago
- RE3: State Entropy Maximization with Random Encoders for Efficient Exploration☆69Jul 29, 2021Updated 4 years ago
- ☆60Apr 16, 2023Updated 2 years ago
- Official code for ACT: Empowering Decision Transformer with Dynamic Programming via Advantage Conditioning (AAAI'24)☆17Feb 10, 2024Updated 2 years ago
- Code for the paper "What Makes Better Augmentation Strategies? Augment Difficult but Not too Different" (ICLR 22)☆12Aug 28, 2023Updated 2 years ago
- Official PyTorch implementation of Scalable Neural Video Representations with Learnable Positional Features (NeurIPS 2022).☆78Apr 3, 2024Updated last year
- Trajectory-wise Multiple Choice Learning for Dynamics Generalization in Reinforcement Learning (NeurIPS 2020)☆39Oct 27, 2020Updated 5 years ago
- ☆18Jun 8, 2023Updated 2 years ago
- CaDM: Context-aware Dynamics Model for Generalization in Model-based Reinforcement Learning☆63May 20, 2020Updated 5 years ago
- ☆322Jan 23, 2022Updated 4 years ago
- A collection of reference environments for offline reinforcement learning☆1,663Nov 18, 2024Updated last year
- ☆53Nov 10, 2022Updated 3 years ago
- ☆17Mar 2, 2023Updated 3 years ago
- ☆60Feb 3, 2023Updated 3 years ago
- Online Decision Transformer☆275Jan 22, 2024Updated 2 years ago
- [NeurIPS 2022] Official codebase for "Meta-Reward-Net: Implicitly Differentiable Reward Learning for Preference-based Reinforcement Learn…☆26Feb 15, 2025Updated last year
- Extreme Q-Learning: Max Entropy RL without Entropy☆87Feb 14, 2023Updated 3 years ago
- Reinforcement Learning via Supervised Learning☆72May 16, 2022Updated 3 years ago
- Reproduction of OpenAI and DeepMind's "Deep Reinforcement Learning from Human Preferences"☆335Nov 29, 2021Updated 4 years ago
- ☆36May 24, 2023Updated 2 years ago
- ☆24Dec 4, 2020Updated 5 years ago
- ☆27Apr 22, 2024Updated last year
- Official codebase for Decision Transformer: Reinforcement Learning via Sequence Modeling.☆2,777Apr 29, 2024Updated last year
- ☆365May 1, 2023Updated 2 years ago
- A Library for Active Preference-based Reward Learning Algorithms☆53Dec 16, 2023Updated 2 years ago
- Online Adaptation of Language Models with a Memory of Amortized Contexts (NeurIPS 2024)☆75Aug 3, 2024Updated last year
- Author's PyTorch implementation of TD7 for online and offline RL☆162Sep 12, 2023Updated 2 years ago