bclarkson-code / TricycleLinks
Autograd to GPT-2 completely from scratch
☆125Updated 4 months ago
Alternatives and similar repositories for Tricycle
Users that are interested in Tricycle are comparing it to the libraries listed below
Sorting:
- A pure NumPy implementation of Mamba.☆222Updated last year
- Visualizing the internal board state of a GPT trained on chess PGN strings, and performing interventions on its internal board state and …☆218Updated last year
- A really tiny autograd engine☆97Updated 7 months ago
- Absolute minimalistic implementation of a GPT-like transformer using only numpy (<650 lines).☆254Updated 2 years ago
- ☆250Updated last year
- Pytorch script hot swap: Change code without unloading your LLM from VRAM☆125Updated 8 months ago
- Simple Transformer in Jax☆140Updated last year
- A tiny version of GPT fully implemented in Python with zero dependencies☆79Updated last year
- a small code base for training large models☆316Updated 8 months ago
- Visualize the intermediate output of Mistral 7B☆381Updated 11 months ago
- in this repository, i'm going to implement increasingly complex llm inference optimizations☆76Updated 7 months ago
- a curated list of data for reasoning ai☆140Updated last year
- Mistral7B playing DOOM☆138Updated last year
- ☆461Updated last month
- Inference of Mamba models in pure C☆196Updated last year
- An implementation of bucketMul LLM inference☆223Updated last year
- This repository contain the simple llama3 implementation in pure jax.☆70Updated 10 months ago
- throwaway GPT inference☆141Updated last year
- Teaching transformers to play chess☆144Updated this week
- Bayesian Optimization as a Coverage Tool for Evaluating LLMs. Accurate evaluation (benchmarking) that's 10 times faster with just a few l…☆287Updated 3 months ago
- noise_step: Training in 1.58b With No Gradient Memory☆220Updated last year
- explore token trajectory trees on instruct and base models☆149Updated 7 months ago
- look how they massacred my boy☆63Updated last year
- Official codebase for the paper "Beyond A* Better Planning with Transformers via Search Dynamics Bootstrapping".☆374Updated last year
- Gradient descent is cool and all, but what if we could delete it?☆104Updated 4 months ago
- ☆115Updated 11 months ago
- run paligemma in real time☆133Updated last year
- Documented and Unit Tested educational Deep Learning framework with Autograd from scratch.☆122Updated last year
- Code to train and evaluate Neural Attention Memory Models to obtain universally-applicable memory systems for transformers.☆345Updated last year
- A BERT that you can train on a (gaming) laptop.☆210Updated 2 years ago