Ki6an / fastT5
⚡ boost inference speed of T5 models by 5x & reduce the model size by 3x.
☆573Updated last year
Alternatives and similar repositories for fastT5:
Users that are interested in fastT5 are comparing it to the libraries listed below
- Summarization, translation, sentiment-analysis, text-generation and more at blazing speed using a T5 version implemented in ONNX.☆252Updated 2 years ago
- ☆496Updated last year
- Efficient, scalable and enterprise-grade CPU/GPU inference server for 🤗 Hugging Face transformer models 🚀☆1,680Updated 4 months ago
- An efficient implementation of the popular sequence models for text generation, summarization, and translation tasks. https://arxiv.org/p…☆430Updated 2 years ago
- Repository for the paper "Optimal Subarchitecture Extraction for BERT"☆471Updated 2 years ago
- Language model fine-tuning on NER with an easy interface and cross-domain evaluation. "T-NER: An All-Round Python Library for Transformer…☆384Updated last year
- [ACL 2021] Learning Dense Representations of Phrases at Scale; EMNLP'2021: Phrase Retrieval Learns Passage Retrieval, Too https://arxiv.o…☆604Updated 2 years ago
- FastFormers - highly efficient transformer models for NLU☆704Updated last year
- Tools to download and cleanup Common Crawl data☆986Updated last year
- NeuSpell: A Neural Spelling Correction Toolkit☆689Updated last year
- This repository contains the code for "Generating Datasets with Pretrained Language Models".☆188Updated 3 years ago
- BLEURT is a metric for Natural Language Generation based on transfer learning.☆718Updated last year
- Python library & examples for Masked Language Model Scoring (ACL 2020)☆340Updated 2 years ago
- Autoregressive Entity Retrieval☆781Updated last year
- XtremeDistil framework for distilling/compressing massive multilingual neural network models to tiny and efficient models for AI at scale☆154Updated last year
- Parallelformers: An Efficient Model Parallelization Toolkit for Deployment☆785Updated last year
- Fast BPE☆666Updated 8 months ago
- Prune a model while finetuning or training.☆399Updated 2 years ago
- Models to perform neural summarization (extractive and abstractive) using machine learning transformers and a tool to convert abstractive…☆429Updated last year
- ICML'2022: NLP From Scratch Without Large-Scale Pretraining: A Simple and Efficient Framework☆259Updated last year
- Accelerated NLP pipelines for fast inference on CPU. Built with Transformers and ONNX runtime.☆126Updated 4 years ago
- Pretrain and finetune ELECTRA with fastai and huggingface. (Results of the paper replicated !)☆326Updated last year
- Fast Inference Solutions for BLOOM☆563Updated 4 months ago
- SentAugment is a data augmentation technique for NLP that retrieves similar sentences from a large bank of sentences. It can be used in c…☆362Updated 3 years ago
- UnifiedQA: Crossing Format Boundaries With a Single QA System☆432Updated 2 years ago
- ☆344Updated 3 years ago
- This dataset contains synthetic training data for grammatical error correction. The corpus is generated by corrupting clean sentences fro…☆160Updated 5 months ago
- Fast + Non-Autoregressive Grammatical Error Correction using BERT. Code and Pre-trained models for paper "Parallel Iterative Edit Models …☆231Updated last year
- Efficient Attention for Long Sequence Processing☆92Updated last year
- simpleT5 is built on top of PyTorch-lightning⚡️ and Transformers🤗 that lets you quickly train your T5 models.☆394Updated last year