yifding / hetseq
HetSeq: Distributed GPU Training on Heterogeneous Infrastructure
☆106Updated last year
Related projects ⓘ
Alternatives and complementary repositories for hetseq
- A case study of efficient training of large language models using commodity hardware.☆68Updated 2 years ago
- Functional deep learning☆106Updated last year
- Check if you have training samples in your test set☆64Updated 2 years ago
- ☆155Updated 4 years ago
- Implementation of Feedback Transformer in Pytorch☆104Updated 3 years ago
- ☆101Updated 3 years ago
- A 🤗-style implementation of BERT using lambda layers instead of self-attention☆70Updated 4 years ago
- The official repository for our paper "The Devil is in the Detail: Simple Tricks Improve Systematic Generalization of Transformers". We s…☆66Updated last year
- A collection of Models, Datasets, DataModules, Callbacks, Metrics, Losses and Loggers to better integrate pytorch-lightning with transfor…☆47Updated last year
- GPT, but made only out of MLPs☆86Updated 3 years ago
- Code for the Shortformer model, from the ACL 2021 paper by Ofir Press, Noah A. Smith and Mike Lewis.☆145Updated 3 years ago
- My implementation of DeepMind's Perceiver☆63Updated 3 years ago
- A collection of code snippets for my PyTorch Lightning projects☆107Updated 3 years ago
- Official code for "Distributed Deep Learning in Open Collaborations" (NeurIPS 2021)☆116Updated 2 years ago
- Swarm training framework using Haiku + JAX + Ray for layer parallel transformer language models on unreliable, heterogeneous nodes☆237Updated last year
- Python Research Framework☆107Updated 2 years ago
- Implementation of the specific Transformer architecture from PaLM - Scaling Language Modeling with Pathways - in Jax (Equinox framework)☆185Updated 2 years ago
- ☆38Updated last year
- ☆64Updated 4 years ago
- Implicit MLE: Backpropagating Through Discrete Exponential Family Distributions☆258Updated last year
- Simple and efficient RevNet-Library for PyTorch with XLA and DeepSpeed support and parameter offload☆124Updated 2 years ago
- Trains Transformer model variants. Data isn't shuffled between batches.☆141Updated 2 years ago
- Training Transformer-XL on 128 GPUs☆140Updated 4 years ago
- Parallel data preprocessing for NLP and ML.☆33Updated 2 weeks ago
- PyTorch implementation of L2L execution algorithm☆106Updated last year
- Named tensors with first-class dimensions for PyTorch☆322Updated last year
- A GPT, made only of MLPs, in Jax☆55Updated 3 years ago
- Docs☆143Updated last month
- diagNNose is a Python library that facilitates a broad set of tools for analysing hidden activations of neural models.☆81Updated last year
- Unit Testing for pytorch, based on mltest☆311Updated 4 years ago