dpressel / mint
MinT: Minimal Transformer Library and Tutorials
☆251Updated 2 years ago
Alternatives and similar repositories for mint:
Users that are interested in mint are comparing it to the libraries listed below
- A pure-functional implementation of a machine learning transformer model in Python/JAX☆175Updated 2 years ago
- A library to inspect and extract intermediate layers of PyTorch models.☆470Updated 2 years ago
- Annotations of the interesting ML papers I read☆224Updated last week
- An interactive exploration of Transformer programming.☆255Updated last year
- Check if you have training samples in your test set☆64Updated 2 years ago
- Module 0 - Fundamentals☆101Updated 4 months ago
- All about the fundamental blocks of TF and JAX!☆272Updated 3 years ago
- Puzzles for exploring transformers☆331Updated last year
- All about the fundamentals and working of Diffusion Models☆152Updated 2 years ago
- FasterAI: Prune and Distill your models with FastAI and PyTorch☆246Updated last month
- Implementation of the specific Transformer architecture from PaLM - Scaling Language Modeling with Pathways - in Jax (Equinox framework)☆184Updated 2 years ago
- Named tensors with first-class dimensions for PyTorch☆322Updated last year
- 100 exercises to learn JAX☆571Updated 2 years ago
- For optimization algorithm research and development.☆484Updated this week
- ☆413Updated 2 months ago
- The "tl;dr" on a few notable transformer papers (pre-2022).☆190Updated 2 years ago
- ☆335Updated 9 months ago
- Host repository for the "Reproducible Deep Learning" PhD course☆403Updated 2 years ago
- Seminar on Large Language Models (COMP790-101 at UNC Chapel Hill, Fall 2022)☆309Updated 2 years ago
- Rax is a Learning-to-Rank library written in JAX.☆324Updated 2 weeks ago
- Resources from the EleutherAI Math Reading Group☆52Updated last month
- Recipes are a standard, well supported set of blueprints for machine learning engineers to rapidly train models using the latest research…☆301Updated this week
- Implementation of the Adan (ADAptive Nesterov momentum algorithm) Optimizer in Pytorch☆250Updated 2 years ago
- ☆197Updated 2 years ago
- A walkthrough of transformer architecture code☆323Updated 10 months ago
- A Jax-based library for designing and training transformer models from scratch.☆280Updated 4 months ago
- A lightweight library designed to accelerate the process of training PyTorch models by providing a minimal, but extensible training loop …☆178Updated 3 weeks ago
- NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day☆253Updated last year