betacord / PSILinks
☆10Updated 7 months ago
Alternatives and similar repositories for PSI
Users that are interested in PSI are comparing it to the libraries listed below
Sorting:
- Kick-off repository for starting with Kaggle!☆12Updated 6 months ago
- Efficient optimizers☆226Updated this week
- supporting pytorch FSDP for optimizers☆82Updated 6 months ago
- Train VAE like a boss☆281Updated 8 months ago
- ☆47Updated 4 months ago
- Focused on fast experimentation and simplicity☆75Updated 6 months ago
- A repository for log-time feedforward networks☆222Updated last year
- A Compressed Stable Diffusion for Efficient Text-to-Image Generation [ECCV'24]☆288Updated 11 months ago
- My take on Flow Matching☆64Updated 5 months ago
- Implementation of the Llama architecture with RLHF + Q-learning☆165Updated 4 months ago
- WIP☆93Updated 10 months ago
- CIFAR-10 speedruns: 94% in 2.6 seconds and 96% in 27 seconds☆252Updated 3 months ago
- ☆133Updated 10 months ago
- This repo implements Diffusion Transformers(DiT) in PyTorch and provides training and inference code on CelebHQ dataset☆39Updated 5 months ago
- Annotated version of the Mamba paper☆485Updated last year
- Re-implementation of Adversarial Diffusion Distillation by AMD☆89Updated 5 months ago
- A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.☆558Updated last week
- When it comes to optimizers, it's always better to be safe than sorry☆244Updated 2 months ago
- ☆190Updated 6 months ago
- neosr is an open-source framework for training super-resolution models.☆261Updated 3 weeks ago
- Simple, minimal implementation of the Mamba SSM in one pytorch file. Using logcumsumexp (Heisen sequence).☆119Updated 8 months ago
- ☆30Updated 8 months ago
- ☆78Updated 11 months ago
- Huggingface-compatible SDXL Unet implementation that is readily hackable☆424Updated last year
- Minimalistic, extremely fast, and hackable researcher's toolbench for GPT models in 307 lines of code. Reaches <3.8 validation loss on wi…☆345Updated 11 months ago
- Training small GPT-2 style models using Kolmogorov-Arnold networks.☆119Updated last year
- ☆12Updated 10 months ago
- Official repository for the paper "Grokfast: Accelerated Grokking by Amplifying Slow Gradients"☆555Updated last year
- Repo for "LoLCATs: On Low-Rank Linearizing of Large Language Models"☆239Updated 4 months ago
- Automated Identification of Redundant Layer Blocks for Pruning in Large Language Models☆238Updated last year