drubinstein / pokemonred_puffer
☆123Updated this week
Alternatives and similar repositories for pokemonred_puffer:
Users that are interested in pokemonred_puffer are comparing it to the libraries listed below
- Visualizing the internal board state of a GPT trained on chess PGN strings, and performing interventions on its internal board state and …☆201Updated 3 months ago
- LLM verified with Monte Carlo Tree Search☆270Updated last month
- Official codebase for the paper "Beyond A* Better Planning with Transformers via Search Dynamics Bootstrapping".☆362Updated 9 months ago
- Mistral7B playing DOOM☆130Updated 8 months ago
- Grandmaster-Level Chess Without Search☆557Updated 2 months ago
- Dead Simple LLM Abliteration☆207Updated 3 weeks ago
- ☆242Updated 11 months ago
- An implementation of bucketMul LLM inference☆215Updated 8 months ago
- The history files when recording human interaction while solving ARC tasks☆97Updated this week
- Bayesian Optimization as a Coverage Tool for Evaluating LLMs. Accurate evaluation (benchmarking) that's 10 times faster with just a few l…☆279Updated 3 weeks ago
- Autograd to GPT-2 completely from scratch☆111Updated this week
- DiscoGrad - automatically differentiate across conditional branches in C++ programs☆203Updated 6 months ago
- Cost aware hyperparameter tuning algorithm☆147Updated 8 months ago
- Grow virtual creatures in static and physics simulated environments.☆52Updated 11 months ago
- Absolute minimalistic implementation of a GPT-like transformer using only numpy (<650 lines).☆250Updated last year
- LeanRL is a fork of CleanRL, where selected PyTorch scripts optimized for performance using compile and cudagraphs.☆520Updated 4 months ago
- Diffusion on syntax trees for program synthesis☆446Updated 8 months ago
- Teaching transformers to play chess☆118Updated last month
- Docker-based inference engine for AMD GPUs☆229Updated 5 months ago
- Run and explore Llama models locally with minimal dependencies on CPU☆191Updated 5 months ago
- A pure NumPy implementation of Mamba.☆219Updated 8 months ago
- Implementation of Q-Transformer, Scalable Offline Reinforcement Learning via Autoregressive Q-Functions, out of Google Deepmind☆366Updated last month
- a curated list of data for reasoning ai☆130Updated 7 months ago
- Code to train and evaluate Neural Attention Memory Models to obtain universally-applicable memory systems for transformers.☆297Updated 4 months ago
- Open weights language model from Google DeepMind, based on Griffin.☆623Updated 3 weeks ago