rasbt / pytorch-memory-optimLinks
This code repository contains the code used for my "Optimizing Memory Usage for Training LLMs and Vision Transformers in PyTorch" blog post.
☆92Updated 2 years ago
Alternatives and similar repositories for pytorch-memory-optim
Users that are interested in pytorch-memory-optim are comparing it to the libraries listed below
Sorting:
- ☆162Updated last year
- Collection of autoregressive model implementation☆86Updated 3 months ago
- A place to store reusable transformer components of my own creation or found on the interwebs☆59Updated last week
- ML/DL Math and Method notes☆62Updated last year
- experiments with inference on llama☆104Updated last year
- ☆83Updated last year
- CUDA and Triton implementations of Flash Attention with SoftmaxN.☆71Updated last year
- ☆48Updated 11 months ago
- ☆88Updated last year
- ☆87Updated last year
- Code for NeurIPS LLM Efficiency Challenge☆59Updated last year
- A really tiny autograd engine☆95Updated 2 months ago
- Various transformers for FSDP research☆37Updated 2 years ago
- 👷 Build compute kernels☆87Updated this week
- ☆81Updated last year
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT training☆130Updated last year
- An introduction to LLM Sampling☆79Updated 7 months ago
- Experiment of using Tangent to autodiff triton☆79Updated last year
- Implementation of the Llama architecture with RLHF + Q-learning☆166Updated 6 months ago
- A miniture AI training framework for PyTorch☆41Updated 6 months ago
- ☆206Updated 5 months ago
- Load compute kernels from the Hub☆220Updated this week
- LoRA and DoRA from Scratch Implementations☆207Updated last year
- Google TPU optimizations for transformers models☆117Updated 6 months ago
- ☆93Updated last year
- Supercharge huggingface transformers with model parallelism.☆77Updated 2 weeks ago
- Advanced Ultra-Low Bitrate Compression Techniques for the LLaMA Family of LLMs☆110Updated last year
- Triton Implementation of HyperAttention Algorithm☆48Updated last year
- Minimal example scripts of the Hugging Face Trainer, focused on staying under 150 lines☆197Updated last year
- A set of scripts and notebooks on LLM finetunning and dataset creation☆110Updated 10 months ago