PABannier / nanograd
A lightweight deep learning framework
☆32Updated 3 years ago
Related projects ⓘ
Alternatives and complementary repositories for nanograd
- Provides symbolic API for model creation in PyTorch.☆62Updated last month
- Neural Networks for JAX☆83Updated last month
- Scripts to prep PC for development use after OS installs☆37Updated this week
- A pure-functional implementation of a machine learning transformer model in Python/JAX☆174Updated 2 years ago
- Automatic gradient descent☆207Updated last year
- Resources from the EleutherAI Math Reading Group☆50Updated last month
- ☆82Updated 8 months ago
- Run PyTorch in JAX. 🤝☆199Updated last year
- A miniscule implementation of reverse mode auto-differentiation☆27Updated 3 years ago
- Train vision models using JAX and 🤗 transformers☆95Updated 2 weeks ago
- ☆116Updated this week
- ☆197Updated 3 months ago
- A Pytree Module system for Deep Learning in JAX☆214Updated last year
- ☆58Updated 2 years ago
- Unofficial JAX implementations of deep learning research papers☆151Updated 2 years ago
- ☆104Updated this week
- Explorations into the proposal from the paper "Grokfast, Accelerated Grokking by Amplifying Slow Gradients"☆84Updated 2 months ago
- Image augmentation library for Jax☆37Updated 7 months ago
- ☆72Updated 4 months ago
- All about the fundamentals and working of Diffusion Models☆152Updated last year
- Serialize JAX, Flax, Haiku, or Objax model params with 🤗`safetensors`☆42Updated 5 months ago
- Scalable neural net training via automatic normalization in the modular norm.☆118Updated 2 months ago
- Multidimensional indexing for tensors☆112Updated last year
- Solve puzzles. Learn CUDA.☆60Updated 10 months ago
- PyTorch interface for TrueGrad Optimizers☆39Updated last year
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT training☆112Updated 6 months ago
- Functional local implementations of main model parallelism approaches☆95Updated last year
- HomebrewNLP in JAX flavour for maintable TPU-Training☆46Updated 9 months ago
- 94% on CIFAR-10 in 2.67 seconds 💨 96% in 27 seconds☆168Updated 3 weeks ago