minitorch / Module-0
Module 0 - Fundamentals
☆102Updated 8 months ago
Alternatives and similar repositories for Module-0:
Users that are interested in Module-0 are comparing it to the libraries listed below
- MinT: Minimal Transformer Library and Tutorials☆254Updated 2 years ago
- Docs☆143Updated 5 months ago
- Resources from the EleutherAI Math Reading Group☆53Updated 2 months ago
- Check if you have training samples in your test set☆64Updated 2 years ago
- A pure-functional implementation of a machine learning transformer model in Python/JAX☆177Updated this week
- A case study of efficient training of large language models using commodity hardware.☆69Updated 2 years ago
- ☆103Updated 4 years ago
- ☆166Updated last year
- Transformer Grammars: Augmenting Transformer Language Models with Syntactic Inductive Biases at Scale, TACL (2022)☆125Updated 6 months ago
- Named tensors with first-class dimensions for PyTorch☆320Updated last year
- HetSeq: Distributed GPU Training on Heterogeneous Infrastructure☆106Updated last year
- ☆108Updated 2 years ago
- ☆430Updated 6 months ago
- A library to create and manage configuration files, especially for machine learning projects.☆78Updated 3 years ago
- ☆153Updated 4 years ago
- HomebrewNLP in JAX flavour for maintable TPU-Training☆50Updated last year
- Framework-agnostic library for checking array/tensor shapes at runtime.☆46Updated 4 years ago
- Swarm training framework using Haiku + JAX + Ray for layer parallel transformer language models on unreliable, heterogeneous nodes☆238Updated last year
- Functional deep learning☆108Updated 2 years ago
- Python Research Framework☆106Updated 2 years ago
- Implementation of the GBST block from the Charformer paper, in Pytorch☆116Updated 3 years ago
- ☆60Updated 3 years ago
- Puzzles for exploring transformers☆344Updated 2 years ago
- An interactive exploration of Transformer programming.☆262Updated last year
- Functional local implementations of main model parallelism approaches☆95Updated 2 years ago
- LM Pretraining with PyTorch/TPU☆134Updated 5 years ago
- Silly twitter torch implementations.☆46Updated 2 years ago
- Code for the Shortformer model, from the ACL 2021 paper by Ofir Press, Noah A. Smith and Mike Lewis.☆147Updated 3 years ago
- Implicit MLE: Backpropagating Through Discrete Exponential Family Distributions☆258Updated last year
- The Python library with command line tools to interact with Dynabench(https://dynabench.org/), such as uploading models.☆55Updated 2 years ago