ZQZCalin / trainitLinks
☆11Updated 3 months ago
Alternatives and similar repositories for trainit
Users that are interested in trainit are comparing it to the libraries listed below
Sorting:
- ☆33Updated last month
- [ICML 2024] SINGD: KFAC-like Structured Inverse-Free Natural Gradient Descent (http://arxiv.org/abs/2312.05705)☆22Updated 9 months ago
- Pytorch implementation of preconditioned stochastic gradient descent (Kron and affine preconditioner, low-rank approximation precondition…☆179Updated 2 weeks ago
- DoG is SGD's Best Friend: A Parameter-Free Dynamic Step Size Schedule☆63Updated last year
- 🧱 Modula software package☆216Updated 2 weeks ago
- PyTorch linear operators for curvature matrices (Hessian, Fisher/GGN, KFAC, ...)☆40Updated 3 months ago
- IVON optimizer for neural networks based on variational learning.☆70Updated 9 months ago
- Parameter-Free Optimizers for Pytorch☆130Updated last year
- ☆206Updated 8 months ago
- Sketched matrix decompositions for PyTorch☆70Updated this week
- ASDL: Automatic Second-order Differentiation Library for PyTorch☆188Updated 8 months ago
- Agustinus' very opiniated publication-ready plotting library☆67Updated 3 months ago
- Pytorch-like dataloaders for JAX.☆94Updated 2 months ago
- Minimal pretraining script for language modeling in PyTorch. Supporting torch compilation and DDP. It includes a model implementation and…☆32Updated this week
- ☆70Updated 8 months ago
- ☆53Updated 10 months ago
- [ICML 2024] SIRFShampoo: Structured inverse- and root-free Shampoo in PyTorch (https://arxiv.org/abs/2402.03496)☆14Updated 9 months ago
- [TMLR 2022] Curvature access through the generalized Gauss-Newton's low-rank structure: Eigenvalues, eigenvectors, directional derivative…☆17Updated 2 years ago
- LoRA for arbitrary JAX models and functions☆140Updated last year
- Sampling with gradient-based Markov Chain Monte Carlo approaches☆107Updated last year
- ☆32Updated 10 months ago
- Minimal, lightweight JAX implementations of popular models.☆79Updated this week
- ☆16Updated 2 years ago
- ☆115Updated 2 months ago
- diffusionjax is a simple and accessible diffusion models package in JAX☆47Updated 6 months ago
- Lightning-like training API for JAX with Flax☆42Updated 8 months ago
- ☆11Updated 4 years ago
- Maximal Update Parametrization (μP) with Flax & Optax.☆16Updated last year
- Code accompanying our paper "Feature Learning in Infinite-Width Neural Networks" (https://arxiv.org/abs/2011.14522)☆62Updated 4 years ago
- Optimization algorithm which fits a ResNet to CIFAR-10 5x faster than SGD / Adam (with terrible generalization)☆14Updated last year