dpressel / mint
MinT: Minimal Transformer Library and Tutorials
☆248Updated 2 years ago
Related projects ⓘ
Alternatives and complementary repositories for mint
- A pure-functional implementation of a machine learning transformer model in Python/JAX☆175Updated 2 years ago
- Named tensors with first-class dimensions for PyTorch☆322Updated last year
- A library to inspect and extract intermediate layers of PyTorch models.☆470Updated 2 years ago
- Annotations of the interesting ML papers I read☆214Updated last week
- An interactive exploration of Transformer programming.☆246Updated last year
- Module 0 - Fundamentals☆99Updated 2 months ago
- Recipes are a standard, well supported set of blueprints for machine learning engineers to rapidly train models using the latest research…☆294Updated this week
- Puzzles for exploring transformers☆325Updated last year
- Check if you have training samples in your test set☆64Updated 2 years ago
- Seminar on Large Language Models (COMP790-101 at UNC Chapel Hill, Fall 2022)☆308Updated last year
- Highly commented implementations of Transformers in PyTorch☆128Updated last year
- Resources from the EleutherAI Math Reading Group☆51Updated last month
- MLCommons Algorithmic Efficiency is a benchmark and competition measuring neural network training speedups due to algorithmic improvement…☆333Updated 3 weeks ago
- ☆391Updated last month
- Implementation of the specific Transformer architecture from PaLM - Scaling Language Modeling with Pathways - in Jax (Equinox framework)☆185Updated 2 years ago
- FasterAI: Prune and Distill your models with FastAI and PyTorch☆243Updated 3 weeks ago
- NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day☆252Updated last year
- ☆161Updated last year
- All about the fundamentals and working of Diffusion Models☆152Updated last year
- The "tl;dr" on a few notable transformer papers (pre-2022).☆189Updated last year
- A walkthrough of transformer architecture code☆318Updated 9 months ago
- Interpretability for sequence generation models 🐛 🔍☆377Updated last week
- ☆334Updated 7 months ago
- ☆187Updated 2 years ago
- Implementation of Flash Attention in Jax☆196Updated 8 months ago
- An interpreter for RASP as described in the ICML 2021 paper "Thinking Like Transformers"☆284Updated 2 months ago
- Helps you write algorithms in PyTorch that adapt to the available (CUDA) memory☆428Updated 2 months ago
- deep learning with pytorch lightning☆0Updated 3 weeks ago
- For optimization algorithm research and development.☆449Updated this week
- Host repository for the "Reproducible Deep Learning" PhD course☆405Updated 2 years ago