amjadmajid / BabyTorch
BabyTorch is a minimalist deep-learning framework with a similar API to PyTorch. This minimalist design encourages learners explore and understand the underlying algorithms and mechanics of deep learning processes. It is design such that when learners are ready to switch to PyTorch they only need to remove the word `baby`.
☆26Updated last year
Alternatives and similar repositories for BabyTorch
Users that are interested in BabyTorch are comparing it to the libraries listed below
Sorting:
- Implementation of Diffusion Transformer (DiT) in JAX☆275Updated 11 months ago
- ☆109Updated this week
- Latent Program Network (from the "Searching Latent Program Spaces" paper)☆82Updated 2 months ago
- Pytorch implementation of Evolutionary Policy Optimization, from Wang et al. of the Robotics Institute at Carnegie Mellon University☆74Updated this week
- Explorations into whether a transformer with RL can direct a genetic algorithm to converge faster☆68Updated 3 weeks ago
- Efficient World Models with Context-Aware Tokenization. ICML 2024☆98Updated 7 months ago
- Official JAX implementation of xLSTM including fast and efficient training and inference code. 7B model available at https://huggingface.…☆91Updated 4 months ago
- Cost aware hyperparameter tuning algorithm☆152Updated 10 months ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆120Updated last week
- Accelerated minigrid environments with JAX☆135Updated this week
- ☆79Updated 6 months ago
- A simple library for scaling up JAX programs☆134Updated 6 months ago
- Just some miscellaneous utility functions / decorators / modules related to Pytorch and Accelerate to help speed up implementation of new…☆121Updated 9 months ago
- JAX Implementation of Black Forest Labs' Flux.1 family of models☆32Updated 6 months ago
- ☆217Updated 10 months ago
- ☆88Updated last year
- Lightning-like training API for JAX with Flax☆38Updated 5 months ago
- seqax = sequence modeling + JAX☆155Updated last month
- Efficient optimizers☆193Updated this week
- fast + parallel AlphaZero in JAX☆96Updated 4 months ago
- For optimization algorithm research and development.☆513Updated this week
- Solve puzzles. Learn CUDA.☆64Updated last year
- Accelerated First Order Parallel Associative Scan☆182Updated 8 months ago
- Minimal but scalable implementation of large language models in JAX☆34Updated 6 months ago
- Home for "How To Scale Your Model", a short blog-style textbook about scaling LLMs on TPUs☆269Updated 3 weeks ago
- WIP☆93Updated 9 months ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆33Updated last year
- Efficiently discovering algorithms via LLMs with evolutionary search and reinforcement learning.☆78Updated 3 weeks ago
- Flexibly track outputs and grad-outputs of torch.nn.Module.☆13Updated last year
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆82Updated last year