betacord / PSILinks
☆12Updated last month
Alternatives and similar repositories for PSI
Users that are interested in PSI are comparing it to the libraries listed below
Sorting:
- Kick-off repository for starting with Kaggle!☆12Updated 11 months ago
- ☆27Updated 10 months ago
- UNet diffusion model in pure CUDA☆651Updated last year
- Efficient optimizers☆276Updated 3 weeks ago
- Minimal implementation of scalable rectified flow transformers, based on SD3's approach☆613Updated last year
- The AdEMAMix Optimizer: Better, Faster, Older.☆186Updated last year
- Implementation of Diffusion Transformer (DiT) in JAX☆293Updated last year
- Official repository for the paper "Grokfast: Accelerated Grokking by Amplifying Slow Gradients"☆564Updated last year
- supporting pytorch FSDP for optimizers☆83Updated 11 months ago
- Quick implementation of nGPT, learning entirely on the hypersphere, from NvidiaAI☆291Updated 5 months ago
- Code for Adam-mini: Use Fewer Learning Rates To Gain More https://arxiv.org/abs/2406.16793☆440Updated 5 months ago
- Implementation of 💍 Ring Attention, from Liu et al. at Berkeley AI, in Pytorch☆542Updated 5 months ago
- A simple implimentation of Bayesian Flow Networks (BFN)☆240Updated last year
- Simple Byte pair Encoding mechanism used for tokenization process . written purely in C☆139Updated 11 months ago
- A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.☆580Updated 2 months ago
- Train VAE like a boss☆297Updated last year
- PyTorch implementation of Infini-Transformer from "Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention…☆292Updated last year
- An extension of the nanoGPT repository for training small MOE models.☆207Updated 8 months ago
- ☆215Updated 10 months ago
- Getting crystal-like representations with harmonic loss☆192Updated 7 months ago
- This repo implements Diffusion Transformers(DiT) in PyTorch and provides training and inference code on CelebHQ dataset☆50Updated 10 months ago
- Annotated version of the Mamba paper☆490Updated last year
- This repository's goal is to precompile all past presentations of the Huggingface reading group☆48Updated last year
- Repo for "Monarch Mixer: A Simple Sub-Quadratic GEMM-Based Architecture"☆560Updated 10 months ago
- Text to Image Latent Diffusion using a Transformer core☆212Updated last year
- CIFAR-10 speedruns: 94% in 2.6 seconds and 96% in 27 seconds☆321Updated 3 months ago
- [ICLR 2025] Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling☆923Updated last week
- EleutherAI ML Performance reading group repository (slides, meeting recordings, annotated papers)☆21Updated last week
- Muon is an optimizer for hidden layers in neural networks☆1,960Updated 3 months ago
- Normalized Transformer (nGPT)☆192Updated 11 months ago