amjadmajid / BabyTorchLinks
BabyTorch is a minimalist deep-learning framework with a similar API to PyTorch. This minimalist design encourages learners explore and understand the underlying algorithms and mechanics of deep learning processes. It is design such that when learners are ready to switch to PyTorch they only need to remove the word `baby`.
☆26Updated 7 months ago
Alternatives and similar repositories for BabyTorch
Users that are interested in BabyTorch are comparing it to the libraries listed below
Sorting:
- Implementation of Diffusion Transformer (DiT) in JAX☆300Updated last year
- CIFAR-10 speedruns: 94% in 2.6 seconds and 96% in 27 seconds☆340Updated last month
- Minimal yet performant LLM examples in pure JAX☆225Updated last week
- Cost aware hyperparameter tuning algorithm☆177Updated last year
- ☆287Updated last year
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆181Updated 6 months ago
- ☆212Updated last year
- Solve puzzles. Learn CUDA.☆63Updated 2 years ago
- Latent Program Network (from the "Searching Latent Program Spaces" paper)☆107Updated last month
- ☆122Updated 6 months ago
- For optimization algorithm research and development.☆556Updated 3 weeks ago
- Pytorch implementation of Evolutionary Policy Optimization, from Wang et al. of the Robotics Institute at Carnegie Mellon University☆103Updated 3 months ago
- The AdEMAMix Optimizer: Better, Faster, Older.☆186Updated last year
- A repository to unravel the language of GPUs, making their kernel conversations easy to understand☆195Updated 7 months ago
- A simple library for scaling up JAX programs☆144Updated 2 months ago
- ☆234Updated last year
- Flax (Jax) implementation of DeepSeek-R1-Distill-Qwen-1.5B with weights ported from Hugging Face.☆26Updated 10 months ago
- ☆131Updated last month
- Custom triton kernels for training Karpathy's nanoGPT.☆19Updated last year
- Synchronized Curriculum Learning for RL Agents☆117Updated 2 months ago
- Official JAX implementation of xLSTM including fast and efficient training and inference code. 7B model available at https://huggingface.…☆105Updated last year
- A zero-to-one guide on scaling modern transformers with n-dimensional parallelism.☆112Updated last week
- 🧱 Modula software package☆322Updated 4 months ago
- Minimal but scalable implementation of large language models in JAX☆35Updated last month
- Distributed pretraining of large language models (LLMs) on cloud TPU slices, with Jax and Equinox.☆24Updated last year
- ☆225Updated last month
- seqax = sequence modeling + JAX☆169Updated 5 months ago
- Accelerate, Optimize performance with streamlined training and serving options with JAX.☆328Updated this week
- An implementation of PSGD Kron second-order optimizer for PyTorch☆97Updated 5 months ago
- supporting pytorch FSDP for optimizers☆84Updated last year