sgrvinod / chess-transformers
Teaching transformers to play chess
☆120Updated 2 months ago
Alternatives and similar repositories for chess-transformers:
Users that are interested in chess-transformers are comparing it to the libraries listed below
- Visualizing the internal board state of a GPT trained on chess PGN strings, and performing interventions on its internal board state and …☆204Updated 4 months ago
- Grandmaster-Level Chess Without Search☆567Updated 3 months ago
- ☆134Updated last week
- A pure NumPy implementation of Mamba.☆222Updated 9 months ago
- Autograd to GPT-2 completely from scratch☆112Updated this week
- Implementation snake game based on Diffusion model☆90Updated 3 months ago
- Helpers and such for working with Lambda Cloud☆51Updated last year
- The history files when recording human interaction while solving ARC tasks☆106Updated this week
- Highly commented implementations of Transformers in PyTorch☆135Updated last year
- A complete end-to-end pipeline for LLM interpretability with sparse autoencoders (SAEs) using Llama 3.2, written in pure PyTorch and full…☆606Updated 3 weeks ago
- A repository for training nanogpt-based Chess playing language models.☆24Updated 11 months ago
- Minimal example scripts of the Hugging Face Trainer, focused on staying under 150 lines☆198Updated 11 months ago
- A multi-player tournament benchmark that tests LLMs in social reasoning, strategy, and deception. Players engage in public and private co…☆246Updated last week
- ☆242Updated last year
- Alice in Wonderland code base for experiments and raw experiments data☆129Updated 2 months ago
- Simplified implementation of UMAP like dimensionality reduction algorithm☆48Updated 4 months ago
- Benchmark LLM reasoning capability by solving chess puzzles.☆73Updated 10 months ago
- Documented and Unit Tested educational Deep Learning framework with Autograd from scratch.☆111Updated last year
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.☆82Updated last year
- MiniHF is an inference, human preference data collection, and fine-tuning tool for local language models. It is intended to help the user…☆169Updated last week
- Simple Transformer in Jax☆136Updated 9 months ago
- Full finetuning of large language models without large memory requirements☆94Updated last year
- run paligemma in real time☆131Updated 10 months ago
- Absolute minimalistic implementation of a GPT-like transformer using only numpy (<650 lines).☆249Updated last year
- ☆27Updated 9 months ago
- ☆150Updated 8 months ago
- NanoGPT-speedrunning for the poor T4 enjoyers☆60Updated last week
- My writings about ARC (Abstraction and Reasoning Corpus)☆76Updated last week
- Mistral7B playing DOOM☆130Updated 9 months ago
- an open source reproduction of NVIDIA's nGPT (Normalized Transformer with Representation Learning on the Hypersphere)☆95Updated last month