Lightning-Universe / lightning-transformersLinks
Flexible components pairing š¤ Transformers with Pytorch Lightning
ā609Updated 2 years ago
Alternatives and similar repositories for lightning-transformers
Users that are interested in lightning-transformers are comparing it to the libraries listed below
Sorting:
- Library for 8-bit optimizers and quantization routines.ā715Updated 2 years ago
- Pretrain and finetune ELECTRA with fastai and huggingface. (Results of the paper replicated !)ā330Updated last year
- FastFormers - highly efficient transformer models for NLUā705Updated 3 months ago
- Understanding the Difficulty of Training Transformersā329Updated 3 years ago
- A simple and working implementation of Electra, the fastest way to pretrain language models from scratch, in Pytorchā227Updated 2 years ago
- Parallelformers: An Efficient Model Parallelization Toolkit for Deploymentā790Updated 2 years ago
- An efficient implementation of the popular sequence models for text generation, summarization, and translation tasks. https://arxiv.org/pā¦ā433Updated 2 years ago
- Pytorch Lightning code guideline for conferencesā1,269Updated last year
- Transformers for Longer Sequencesā615Updated 2 years ago
- Fast Block Sparse Matrices for Pytorchā547Updated 4 years ago
- Toolbox of models, callbacks, and datasets for AI/ML researchers.ā1,730Updated 3 weeks ago
- SentAugment is a data augmentation technique for NLP that retrieves similar sentences from a large bank of sentences. It can be used in cā¦ā362Updated 3 years ago
- Repository containing code for "How to Train BERT with an Academic Budget" paperā313Updated last year
- Configuration classes enabling type-safe PyTorch configuration for Hydra appsā218Updated 2 years ago
- Fast, general, and tested differentiable structured prediction in PyTorchā1,113Updated 3 years ago
- Implementation of https://arxiv.org/abs/1904.00962ā377Updated 4 years ago
- Repository for the paper "Optimal Subarchitecture Extraction for BERT"ā473Updated 3 years ago
- A Visual Analysis Tool to Explore Learned Representations in Transformers Modelsā595Updated last year
- Organize your experiments into discrete steps that can be cached and reused throughout the lifetime of your research project.ā561Updated last year
- Prune a model while finetuning or training.ā403Updated 3 years ago
- Run Effective Large Batch Contrastive Learning Beyond GPU/TPU Memory Constraintā395Updated last year
- NL-Augmenter š¦ ā š A Collaborative Repository of Natural Language Transformationsā787Updated last year
- Sinkhorn Transformer - Practical implementation of Sparse Sinkhorn Attentionā264Updated 3 years ago
- Interpretable Evaluation for AI Systemsā366Updated 2 years ago
- Code for the ALiBi method for transformer language models (ICLR 2022)ā535Updated last year
- Cockpit: A Practical Debugging Tool for Training Deep Neural Networksā480Updated 2 years ago
- A library to inspect and extract intermediate layers of PyTorch models.ā473Updated 3 years ago
- Pytorch library for fast transformer implementationsā1,718Updated 2 years ago
- Fully featured implementation of Routing Transformerā295Updated 3 years ago
- Long Range Arena for Benchmarking Efficient Transformersā757Updated last year