pytorch-tpu / examples
This repository contains example code to build models on TPUs
☆30Updated last year
Related projects ⓘ
Alternatives and complementary repositories for examples
- A case study of efficient training of large language models using commodity hardware.☆68Updated 2 years ago
- Standalone pre-training recipe with JAX+Flax☆31Updated last year
- HomebrewNLP in JAX flavour for maintable TPU-Training☆46Updated 9 months ago
- Code for scaling Transformers☆26Updated 3 years ago
- PyTorch implementation of GLOM☆21Updated 2 years ago
- A collection of Models, Datasets, DataModules, Callbacks, Metrics, Losses and Loggers to better integrate pytorch-lightning with transfor…☆47Updated last year
- Various transformers for FSDP research☆33Updated last year
- Babysit your preemptible TPUs☆84Updated last year
- ☆86Updated 2 years ago
- Fast, Modern, Memory Efficient, and Low Precision PyTorch Optimizers☆58Updated 3 months ago
- My explorations into editing the knowledge and memories of an attention network☆34Updated last year
- ☆56Updated 2 years ago
- Code for the Shortformer model, from the ACL 2021 paper by Ofir Press, Noah A. Smith and Mike Lewis.☆145Updated 3 years ago
- ☆29Updated 2 weeks ago
- TPU support for the fastai library☆13Updated 3 years ago
- Helper scripts and notes that were used while porting various nlp models☆44Updated 2 years ago
- Dense Passage Retrieval using tensorflow-keras on TPU☆15Updated 3 years ago
- LM Pretraining with PyTorch/TPU☆132Updated 5 years ago
- A GPT, made only of MLPs, in Jax☆55Updated 3 years ago
- ☆12Updated 2 years ago
- Training and evaluation code for the paper "Headless Language Models: Learning without Predicting with Contrastive Weight Tying" (https:/…☆23Updated 6 months ago
- A python library for highly configurable transformers - easing model architecture search and experimentation.☆49Updated 2 years ago
- ☆64Updated 2 years ago
- A library to create and manage configuration files, especially for machine learning projects.☆77Updated 2 years ago
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.☆92Updated last year
- GPT, but made only out of MLPs☆86Updated 3 years ago
- Implementation of N-Grammer, augmenting Transformers with latent n-grams, in Pytorch☆72Updated last year
- A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training☆26Updated 2 years ago
- A lightweight wrapper for PyTorch that provides a simple declarative API for context switching between devices, distributed modes, mixed-…☆66Updated last year