betacord / PSILinks
☆11Updated 2 months ago
Alternatives and similar repositories for PSI
Users that are interested in PSI are comparing it to the libraries listed below
Sorting:
- Kick-off repository for starting with Kaggle!☆12Updated last year
- ☆27Updated 11 months ago
- Efficient optimizers☆277Updated last month
- Minimal implementation of scalable rectified flow transformers, based on SD3's approach☆624Updated last year
- Code for Adam-mini: Use Fewer Learning Rates To Gain More https://arxiv.org/abs/2406.16793☆445Updated 7 months ago
- ☆227Updated 11 months ago
- UNet diffusion model in pure CUDA☆656Updated last year
- EleutherAI ML Performance reading group repository (slides, meeting recordings, annotated papers)☆23Updated this week
- supporting pytorch FSDP for optimizers☆84Updated last year
- ☆574Updated last year
- End-to-end recipes for optimizing diffusion models with torchao and diffusers (inference and FP8 training).☆388Updated 6 months ago
- Official repository for the paper "Grokfast: Accelerated Grokking by Amplifying Slow Gradients"☆565Updated last year
- A Compressed Stable Diffusion for Efficient Text-to-Image Generation [ECCV'24]☆305Updated last year
- Train VAE like a boss☆307Updated last year
- Text to Image Latent Diffusion using a Transformer core☆216Updated last year
- Annotated version of the Mamba paper☆492Updated last year
- Implementation of Diffusion Transformer (DiT) in JAX☆298Updated last year
- CIFAR-10 speedruns: 94% in 2.6 seconds and 96% in 27 seconds☆335Updated last month
- ☆15Updated 9 months ago
- A curated list of resources for learning and exploring Triton, OpenAI's programming language for writing efficient GPU code.☆441Updated 9 months ago
- Dion optimizer algorithm☆404Updated this week
- Simple, minimal implementation of the Mamba SSM in one pytorch file. Using logcumsumexp (Heisen sequence).☆128Updated last year
- Code for Neurips24 paper: QuaRot, an end-to-end 4-bit inference of large language models.☆466Updated last year
- Best practices & guides on how to write distributed pytorch training code☆552Updated last month
- A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.☆585Updated 4 months ago
- ☆45Updated 6 months ago
- This repo implements Diffusion Transformers(DiT) in PyTorch and provides training and inference code on CelebHQ dataset☆53Updated 11 months ago
- Diffusion Reading Group at EleutherAI☆333Updated 2 years ago
- Helpful tools and examples for working with flex-attention☆1,089Updated this week
- Code for paper: "QuIP: 2-Bit Quantization of Large Language Models With Guarantees"☆391Updated last year