huggingface / picotron_tutorial
☆91Updated this week
Alternatives and similar repositories for picotron_tutorial:
Users that are interested in picotron_tutorial are comparing it to the libraries listed below
- ☆75Updated 6 months ago
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, spars…☆254Updated 3 weeks ago
- ☆47Updated 4 months ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆89Updated last month
- Train, tune, and infer Bamba model☆73Updated this week
- Collection of autoregressive model implementation☆76Updated this week
- ☆138Updated 11 months ago
- This code repository contains the code used for my "Optimizing Memory Usage for Training LLMs and Vision Transformers in PyTorch" blog po…☆87Updated last year
- NanoGPT (124M) quality in 2.67B tokens☆24Updated this week
- Tree Attention: Topology-aware Decoding for Long-Context Attention on GPU clusters☆109Updated last month
- ring-attention experiments☆113Updated 2 months ago
- 🚀 Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flash…☆213Updated this week
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT training☆121Updated 8 months ago
- Fast, Modern, Memory Efficient, and Low Precision PyTorch Optimizers☆77Updated 5 months ago
- ML/DL Math and Method notes☆57Updated last year
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆81Updated last year
- ☆40Updated 11 months ago
- Minimal sharded dataset loaders, decoders, and utils for multi-modal document, image, and text datasets.☆155Updated 9 months ago
- ☆78Updated 8 months ago
- The source code of our work "Prepacking: A Simple Method for Fast Prefilling and Increased Throughput in Large Language Models"☆57Updated 3 months ago
- Prune transformer layers☆67Updated 7 months ago
- PyTorch per step fault tolerance (actively under development)☆209Updated this week
- Best practices & guides on how to write distributed pytorch training code☆329Updated 3 weeks ago
- experiments with inference on llama☆104Updated 7 months ago
- Supercharge huggingface transformers with model parallelism.☆75Updated 3 months ago
- ☆296Updated 6 months ago
- Scalable and Performant Data Loading☆201Updated this week
- CUDA and Triton implementations of Flash Attention with SoftmaxN.☆67Updated 7 months ago
- Google TPU optimizations for transformers models☆84Updated this week
- Pytorch implementation of the PEER block from the paper, Mixture of A Million Experts, by Xu Owen He at Deepmind☆115Updated 4 months ago