microsoft / Lightweight-Low-Resource-NMTLinks
Official code for "Too Brittle To Touch: Comparing the Stability of Quantization and Distillation Towards Developing Lightweight Low-Resource MT Models" to appear in WMT 2022.
☆17Updated last year
Alternatives and similar repositories for Lightweight-Low-Resource-NMT
Users that are interested in Lightweight-Low-Resource-NMT are comparing it to the libraries listed below
Sorting:
- We release the UICaption dataset. The dataset consists of UI images (icons and screenshots) and associated text descriptions. This datase…☆41Updated 2 years ago
- CyBERTron-LM is a project which collects some pre-trained Transformer-based models.☆12Updated 2 years ago
- Fault-aware neural code rankers☆28Updated 2 years ago
- ☆84Updated 2 years ago
- ☆14Updated 2 years ago
- ☆22Updated 2 years ago
- We introduce MKQA, an open-domain question answering evaluation set comprising 10k question-answer pairs aligned across 26 typologically …☆182Updated 3 years ago
- DeFacto - Demonstrations and Feedback for improving factual consistency of text summarization☆29Updated 2 years ago
- A client library for LAION's effort to filter CommonCrawl with CLIP, building a large scale image-text dataset.☆32Updated 2 years ago
- An instruction-based benchmark for text improvements.☆141Updated 2 years ago
- ☆100Updated 2 years ago
- Experiments for "Automatic Calibration and Error Correction for Large Language Models via Pareto Optimal Self-Supervision"☆13Updated last year
- A suite of tools for managing crowdsourcing tasks from the inception through to data packaging for research use.☆312Updated 7 months ago
- Generating Captions via Perceiver-Resampler Cross-Attention Networks☆17Updated 2 years ago
- Research code for pixel-based encoders of language (PIXEL)☆337Updated this week
- BLOOM+1: Adapting BLOOM model to support a new unseen language☆73Updated last year
- Pipeline for pulling and processing online language model pretraining data from the web☆178Updated last year
- A library for preparing data for machine translation research (monolingual preprocessing, bitext mining, etc.) built by the FAIR NLLB te…☆280Updated 5 months ago
- A fast implementation of T5/UL2 in PyTorch using Flash Attention☆105Updated 4 months ago
- ☆66Updated last year
- Tutorial to pretrain & fine-tune a 🤗 Flax T5 model on a TPUv3-8 with GCP☆58Updated 2 years ago
- ☆215Updated 3 weeks ago
- Consists of the largest (10K) human annotated code-switched semantic parsing dataset & 170K generated utterance using the CST5 augmentati…☆39Updated 2 years ago
- This project studies the performance and robustness of language models and task-adaptation methods.☆150Updated last year
- Code used for the creation of OBELICS, an open, massive and curated collection of interleaved image-text web documents, containing 141M d…☆205Updated 10 months ago
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.☆93Updated 2 years ago
- NTREX -- News Test References for MT Evaluation☆84Updated last year
- UNISUMM: Unified Few-shot Summarization with Multi-Task Pre-Training and Prefix-Tuning☆60Updated 2 years ago
- ☆102Updated 7 months ago
- ☆19Updated 2 years ago