Geotrend-research / smaller-transformersLinks
Load What You Need: Smaller Multilingual Transformers for Pytorch and TensorFlow 2.0.
β105Updated 3 years ago
Alternatives and similar repositories for smaller-transformers
Users that are interested in smaller-transformers are comparing it to the libraries listed below
Sorting:
- Tutorial to pretrain & fine-tune a π€ Flax T5 model on a TPUv3-8 with GCPβ58Updated 3 years ago
- XtremeDistil framework for distilling/compressing massive multilingual neural network models to tiny and efficient models for AI at scaleβ157Updated 2 years ago
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.β96Updated 2 years ago
- This repository contains the code for the paper 'PARM: Paragraph Aggregation Retrieval Model for Dense Document-to-Document Retrieval' puβ¦β41Updated 3 years ago
- As good as new. How to successfully recycle English GPT-2 to make models for other languages (ACL Findings 2021)β48Updated 4 years ago
- β75Updated 4 years ago
- Accelerated NLP pipelines for fast inference on CPU. Built with Transformers and ONNX runtime.β127Updated 5 years ago
- State of the art Semantic Sentence Embeddingsβ99Updated 3 years ago
- Fine-tune transformers with pytorch-lightningβ44Updated 3 years ago
- Open source library for few shot NLPβ78Updated 2 years ago
- On Generating Extended Summaries of Long Documentsβ78Updated 4 years ago
- Dual Encoders for State-of-the-art Natural Language Processing.β61Updated 3 years ago
- β101Updated 3 years ago
- Main repository for "CharacterBERT: Reconciling ELMo and BERT for Word-Level Open-Vocabulary Representations From Characters"β200Updated 2 years ago
- Build a dialog dataset from online books in many languagesβ76Updated 3 years ago
- Multilingual abstractive summarization dataset extracted from WikiHow.β97Updated 9 months ago
- Implementation of the paper 'Sentence Bottleneck Autoencoders from Transformer Language Models'β17Updated 3 years ago
- [EMNLP-Findings 2020] Adapting BERT for Word Sense Disambiguation with Gloss Selection Objective and Example Sentencesβ63Updated last year
- Code and models used in "MUSS Multilingual Unsupervised Sentence Simplification by Mining Paraphrases".β100Updated 2 years ago
- Implementation, trained models and result data for the paper "Aspect-based Document Similarity for Research Papers" #COLING2020β63Updated last year
- A repository for our AAAI-2020 Cross-lingual-NER paper. Code will be updated shortly.β47Updated 3 years ago
- This repository contains the code for "BERTRAM: Improved Word Embeddings Have Big Impact on Contextualized Representations".β64Updated 5 years ago
- β68Updated 7 months ago
- A package for fine-tuning Transformers with TPUs, written in Tensorflow2.0+β37Updated 4 years ago
- QED: A Framework and Dataset for Explanations in Question Answeringβ119Updated 4 years ago
- On the Stability of Fine-tuning BERT: Misconceptions, Explanations, and Strong Baselinesβ137Updated 2 years ago
- BERT, RoBERTa fine-tuning over SQuAD Dataset using pytorch-lightningβ‘οΈ, π€-transformers & π€-nlp.β36Updated 2 years ago
- Research code for the paper "How Good is Your Tokenizer? On the Monolingual Performance of Multilingual Language Models"β27Updated 4 years ago
- Training T5 to perform numerical reasoning.β24Updated 4 years ago
- Using business-level retrieval system (BM25) with Python in just a few lines.β31Updated 2 years ago