lucas-maes / nano-simsiamLinks
Minimalistic, hackable PyTorch implementation of SimSiam in ~400 lines. Achieves good performance on ImageNet with ResNet50. Features distributed training, real-time KNN eval, and AMP. Perfect for research prototyping.
☆19Updated 6 months ago
Alternatives and similar repositories for nano-simsiam
Users that are interested in nano-simsiam are comparing it to the libraries listed below
Sorting:
- ☆29Updated 6 months ago
- ☆32Updated last year
- Transformer with Mu-Parameterization, implemented in Jax/Flax. Supports FSDP on TPU pods.☆30Updated last week
- An implementation of the Llama architecture, to instruct and delight☆21Updated last week
- ☆32Updated 8 months ago
- ☆17Updated 9 months ago
- ☆60Updated 3 years ago
- Implementation of Gradient Agreement Filtering, from Chaubard et al. of Stanford, but for single machine microbatches, in Pytorch☆25Updated 4 months ago
- Simple Scalable Discrete Diffusion for text in PyTorch☆33Updated 8 months ago
- ☆32Updated last year
- ☆43Updated last week
- A simple library for scaling up JAX programs☆137Updated 7 months ago
- Minimal but scalable implementation of large language models in JAX☆34Updated 7 months ago
- ☆34Updated 8 months ago
- A simple hypernetwork implementation in jax using haiku.☆23Updated 2 years ago
- ☆27Updated last year
- ICML 2022: Learning Iterative Reasoning through Energy Minimization☆46Updated 2 years ago
- ☆51Updated 11 months ago
- Unofficial but Efficient Implementation of "Mamba: Linear-Time Sequence Modeling with Selective State Spaces" in JAX☆82Updated last year
- ☆13Updated last week
- Learn online intrinsic rewards from LLM feedback☆37Updated 5 months ago
- Code for A General Recipe for Likelihood-free Bayesian Optimization, ICML 2022☆44Updated 2 years ago
- ☆40Updated last year
- JAX implementation of VQVAE/VQGAN autoencoders (+FSQ)☆29Updated last year
- Implementation of PSGD optimizer in JAX☆33Updated 5 months ago
- Personal solutions to the Triton Puzzles☆18Updated 10 months ago
- Code for the paper "Function-Space Learning Rates"☆20Updated this week
- Tensor Parallelism with JAX + Shard Map☆11Updated last year
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆33Updated last year
- Experiment of using Tangent to autodiff triton☆79Updated last year