bclarkson-code / TricycleLinks
Autograd to GPT-2 completely from scratch
☆113Updated last month
Alternatives and similar repositories for Tricycle
Users that are interested in Tricycle are comparing it to the libraries listed below
Sorting:
- A pure NumPy implementation of Mamba.☆223Updated 10 months ago
- Pytorch script hot swap: Change code without unloading your LLM from VRAM☆126Updated last month
- Visualizing the internal board state of a GPT trained on chess PGN strings, and performing interventions on its internal board state and …☆205Updated 6 months ago
- a curated list of data for reasoning ai☆136Updated 9 months ago
- Absolute minimalistic implementation of a GPT-like transformer using only numpy (<650 lines).☆251Updated last year
- ☆243Updated last year
- look how they massacred my boy☆63Updated 7 months ago
- A tiny version of GPT fully implemented in Python with zero dependencies☆70Updated 5 months ago
- Pivotal Token Search☆89Updated 2 weeks ago
- Mistral7B playing DOOM☆131Updated 10 months ago
- This repository contain the simple llama3 implementation in pure jax.☆64Updated 3 months ago
- Simple Transformer in Jax☆137Updated 11 months ago
- Bayesian Optimization as a Coverage Tool for Evaluating LLMs. Accurate evaluation (benchmarking) that's 10 times faster with just a few l…☆284Updated last week
- A really tiny autograd engine☆94Updated last week
- An easy-to-understand framework for LLM samplers that rewind and revise generated tokens☆139Updated 3 months ago
- ☆130Updated 9 months ago
- Teaching transformers to play chess☆125Updated 4 months ago
- NanoGPT-speedrunning for the poor T4 enjoyers☆66Updated last month
- Full finetuning of large language models without large memory requirements☆93Updated last year
- An implementation of bucketMul LLM inference☆217Updated 11 months ago
- a small code base for training large models☆299Updated last month
- run paligemma in real time☆131Updated last year
- Fast parallel LLM inference for MLX☆189Updated 10 months ago
- throwaway GPT inference☆139Updated last year
- Lightweight Llama 3 8B Inference Engine in CUDA C☆46Updated 2 months ago
- R.L. methods and techniques.☆191Updated 6 months ago
- Video+code lecture on building nanoGPT from scratch☆67Updated 11 months ago
- Automated Capability Discovery via Foundation Model Self-Exploration☆49Updated 3 months ago
- in this repository, i'm going to implement increasingly complex llm inference optimizations☆58Updated last week
- A simple MLX implementation for pretraining LLMs on Apple Silicon.☆76Updated last month