DHDev0 / Stochastic-muzeroView external linksLinks
Pytorch Implementation of Stochastic MuZero for gym environment. This algorithm is capable of supporting a wide range of action and observation spaces, including both discrete and continuous variations.
☆75Dec 31, 2025Updated last month
Alternatives and similar repositories for Stochastic-muzero
Users that are interested in Stochastic-muzero are comparing it to the libraries listed below
Sorting:
- Pytorch Implementation of MuZero Unplugged for gym environment. This algorithm is capable of supporting a wide range of action and observ…☆35Jun 25, 2025Updated 7 months ago
- A number of agents (PPO, MuZero) with a Perceiver-based NN architecture that can be trained to achieve goals in nethack/minihack environm…☆43Sep 19, 2022Updated 3 years ago
- A C++ pytorch implementation of MuZero☆40May 1, 2024Updated last year
- A project that provides help for using DeepMind's mctx on gym-style environments.☆64Nov 14, 2024Updated last year
- Pytorch Implementation of MuZero for gym environment. It support any Discrete , Box and Box2D configuration for the action space and obse…☆19Jan 24, 2023Updated 3 years ago
- A clean and easy implementation of MuZero, AlphaZero and Self-Play reinforcement learning algorithms for any game.☆17Oct 15, 2024Updated last year
- [IEEE ToG] MiniZero: An AlphaZero and MuZero Training Framework☆121Updated this week
- Pytorch Implementation of MuZero☆352Jul 23, 2023Updated 2 years ago
- [NeurIPS 2023 Spotlight] LightZero: A Unified Benchmark for Monte Carlo Tree Search in General Sequential Decision Scenarios (awesome MCT…☆1,530Updated this week
- [ICML 2024, Spotlight] EfficientZero V2: Mastering Discrete and Continuous Control with Limited Data☆101Aug 9, 2024Updated last year
- A clean implementation of MuZero and AlphaZero following the AlphaZero General framework. Train and Pit both algorithms against each othe…☆168Mar 28, 2021Updated 4 years ago
- Advantage Alignment Algorithms (ICLR 2025 oral)☆16Apr 7, 2025Updated 10 months ago
- ♟️ Vectorized RL game environments in JAX☆585Mar 6, 2025Updated 11 months ago
- Implementation of some of the Deep Distributional Reinforcement Learning Algorithms.☆25Jun 17, 2025Updated 7 months ago
- Open-source codebase for EfficientZero, from "Mastering Atari Games with Limited Data" at NeurIPS 2021.☆923Dec 20, 2023Updated 2 years ago
- Swarm learning algorithm☆11Jun 2, 2021Updated 4 years ago
- MuZero☆2,766Sep 3, 2024Updated last year
- Deep memory and sequence models in JAX☆23Jan 15, 2026Updated last month
- ☆53Apr 11, 2023Updated 2 years ago
- ☆12Apr 22, 2022Updated 3 years ago
- A high throughput, end-to-end RL library for infinite-horizon tasks.☆23Oct 22, 2025Updated 3 months ago
- ☆14Aug 18, 2023Updated 2 years ago
- Adding Dreamer-v3's implementation tricks to CleanRL's PPO☆14May 19, 2023Updated 2 years ago
- Simple Distributed Reinforcement Learning Framework(シンプルな分散強化学習フレームワーク)☆58Nov 15, 2025Updated 3 months ago
- Code for our TMLR paper "Distributional GFlowNets with Quantile Flows".☆13Feb 14, 2024Updated 2 years ago
- A PyTorch implementation of DeepMind's MuZero agent☆36Dec 1, 2023Updated 2 years ago
- Deep Reinforcement Learning Framework done with PyTorch☆43Mar 12, 2025Updated 11 months ago
- A PyTorch implementation of SEED, originally created by Google Research for TensorFlow 2.☆15Dec 8, 2020Updated 5 years ago
- 🏛️A research-friendly codebase for fast experimentation of single-agent reinforcement learning in JAX • End-to-End JAX RL☆390Oct 29, 2025Updated 3 months ago
- Exploration into the Firefly algorithm in Pytorch☆41Feb 14, 2025Updated last year
- AlphaZero for continuous control tasks☆23Dec 7, 2022Updated 3 years ago
- ☆18Nov 4, 2021Updated 4 years ago
- PyTorch Implementation of the Maximum a Posteriori Policy Optimisation☆79Nov 19, 2022Updated 3 years ago
- ☆46Sep 24, 2024Updated last year
- A method to train DRL model with Tensorflow and Bizhawk.☆25Nov 12, 2019Updated 6 years ago
- ☆22Jan 15, 2026Updated last month
- ☆90Nov 3, 2024Updated last year
- Drop-in environment replacements that make your RL algorithm train faster.☆21Jun 19, 2024Updated last year
- ☆20May 22, 2022Updated 3 years ago