hundredblocks / large-model-parallelismLinks
Functional local implementations of main model parallelism approaches
โ95Updated 2 years ago
Alternatives and similar repositories for large-model-parallelism
Users that are interested in large-model-parallelism are comparing it to the libraries listed below
Sorting:
- Large scale 4D parallelism pre-training for ๐ค transformers in Mixture of Experts *(still work in progress)*โ85Updated last year
- Train very large language models in Jax.โ204Updated last year
- A puzzle to learn about promptingโ131Updated 2 years ago
- Inference code for LLaMA models in JAXโ118Updated last year
- git extension for {collaborative, communal, continual} model developmentโ214Updated 8 months ago
- โ92Updated last year
- JAX implementation of the Llama 2 modelโ219Updated last year
- โ79Updated last year
- Automatic gradient descentโ208Updated 2 years ago
- gzip Predicts Data-dependent Scaling Lawsโ35Updated last year
- โ53Updated last year
- some common Huggingface transformers in maximal update parametrization (ยตP)โ81Updated 3 years ago
- HomebrewNLP in JAX flavour for maintable TPU-Trainingโ50Updated last year
- A lightweight PyTorch implementation of the Transformer-XL architecture proposed by Dai et al. (2019)โ37Updated 2 years ago
- โ61Updated last year
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.โ82Updated last year
- Project 2 (Building Large Language Models) for Stanford CS324: Understanding and Developing Large Language Models (Winter 2022)โ105Updated 2 years ago
- โ143Updated 2 years ago
- ML/DL Math and Method notesโ61Updated last year
- โ61Updated 3 years ago
- Exploring finetuning public checkpoints on filter 8K sequences on Pileโ116Updated 2 years ago
- Reimplementation of the task generation part from the Alpaca paperโ119Updated 2 years ago
- Supercharge huggingface transformers with model parallelism.โ77Updated 9 months ago
- An interactive exploration of Transformer programming.โ265Updated last year
- โ20Updated last year
- Resources from the EleutherAI Math Reading Groupโ53Updated 4 months ago
- A place to store reusable transformer components of my own creation or found on the interwebsโ56Updated last week
- A case study of efficient training of large language models using commodity hardware.โ68Updated 2 years ago
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT trainingโ129Updated last year
- Training and Inference Notebooks for the RedPajama (OpenLlama) modelsโ18Updated 2 years ago