maroxtn / mt5-M2M-comparison
Comparing M2M and mT5 on a rare language pairs, blog post: https://medium.com/@abdessalemboukil/comparing-facebooks-m2m-to-mt5-in-low-resources-translation-english-yoruba-ef56624d2b75
☆15Updated 3 years ago
Alternatives and similar repositories for mt5-M2M-comparison:
Users that are interested in mt5-M2M-comparison are comparing it to the libraries listed below
- ☆38Updated 2 years ago
- The source code of "Language Models are Few-shot Multilingual Learners" (MRL @ EMNLP 2021)☆52Updated 2 years ago
- Using short models to classify long texts☆21Updated last year
- Code for ProtAugment: Unsupervised diverse short-texts paraphrasing for intent detection meta-learning☆21Updated 2 years ago
- A crowdsourced dataset of dialogues grounded in social contexts involving utilization of commonsense.☆78Updated 3 years ago
- Fine-tuned BERT on SQuAd 2.0 Dataset. Applied Knowledge Distillation (KD) and fine-tuned DistilBERT (student) using BERT as the teacher m…☆25Updated 4 years ago
- An example of multilingual machine translation using a pretrained version of mt5 from Hugging Face.☆42Updated 3 years ago
- Benchmarking various Deep Learning models such as BERT, ALBERT, BiLSTMs on the task of sentence entailment using two datasets - MultiNLI …☆28Updated 4 years ago
- pyTorch implementation of Recurrence over BERT (RoBERT) based on this paper https://arxiv.org/abs/1910.10781 and comparison with pyTorch …☆80Updated 2 years ago
- Zero-shot Transfer Learning from English to Arabic☆29Updated 2 years ago
- This repository contains code for the paper "Meet Your Favorite Character: Open-domain Chatbot Mimicking Fictional Characters with only a…☆13Updated 2 years ago
- This is a new metric that can be used to evaluate faithfulness of text generated by LLMs. The work behind this repository can be found he…☆31Updated last year
- ☆30Updated 4 years ago
- Awesome Question Answering☆28Updated 2 years ago
- ARCHIVED. Please use https://docs.adapterhub.ml/huggingface_hub.html || 🔌 A central repository collecting pre-trained adapter modules☆68Updated 8 months ago
- ☆51Updated last year
- ☆34Updated 4 years ago
- Contains notebooks related to various transformers based models for different nlp based tasks☆42Updated last year
- A collection of preprocessed datasets and pretrained models for generating paraphrases.☆29Updated 3 years ago
- Consists of the largest (10K) human annotated code-switched semantic parsing dataset & 170K generated utterance using the CST5 augmentati…☆37Updated 2 years ago
- Vocabulary Trimming (VT) is a model compression technique, which reduces a multilingual LM vocabulary to a target language by deleting ir…☆34Updated 3 months ago
- Finetune GPT2 for text summarization☆15Updated 3 years ago
- BLOOM+1: Adapting BLOOM model to support a new unseen language☆70Updated 11 months ago
- ☆32Updated 2 years ago
- ☆11Updated 2 years ago
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.☆93Updated 2 years ago
- Open source library for few shot NLP☆77Updated last year
- The official code for PRIMERA: Pyramid-based Masked Sentence Pre-training for Multi-document Summarization☆156Updated 2 years ago
- PyTorch implementation of NMT models along with custom tokenizers, models, and datasets☆20Updated 2 years ago