google-deepmind / nanodo
☆197Updated 4 months ago
Related projects ⓘ
Alternatives and complementary repositories for nanodo
- seqax = sequence modeling + JAX☆133Updated 4 months ago
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆516Updated this week
- A simple library for scaling up JAX programs☆127Updated 2 weeks ago
- Cost aware hyperparameter tuning algorithm☆123Updated 4 months ago
- Scalable neural net training via automatic normalization in the modular norm.☆121Updated 3 months ago
- Accelerated First Order Parallel Associative Scan☆163Updated 3 months ago
- A MAD laboratory to improve AI architecture designs 🧪☆95Updated 6 months ago
- ☆224Updated 4 months ago
- LoRA for arbitrary JAX models and functions☆132Updated 8 months ago
- Named Tensors for Legible Deep Learning in JAX☆153Updated this week
- JAX implementation of the Llama 2 model☆210Updated 9 months ago
- Run PyTorch in JAX. 🤝☆200Updated last year
- ☆128Updated this week
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆84Updated last week
- jax-triton contains integrations between JAX and OpenAI Triton☆343Updated 3 weeks ago
- Understand and test language model architectures on synthetic tasks.☆162Updated 6 months ago
- JAX Synergistic Memory Inspector☆164Updated 4 months ago
- ☆303Updated this week
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT training☆113Updated 7 months ago
- ☆105Updated 2 weeks ago
- ☆207Updated 6 months ago
- For optimization algorithm research and development.☆449Updated this week
- JAX-Toolbox☆245Updated this week
- A Jax-based library for designing and training transformer models from scratch.☆276Updated 2 months ago
- Puzzles for exploring transformers☆325Updated last year
- Implementation of Diffusion Transformer (DiT) in JAX☆252Updated 5 months ago
- ☆73Updated 4 months ago
- Annotated version of the Mamba paper☆457Updated 8 months ago
- Inference code for LLaMA models in JAX☆113Updated 6 months ago
- Muon optimizer for neural networks: >30% extra sample efficiency, <3% wallclock overhead☆109Updated last week