microsoft / CyBERTron-LM
CyBERTron-LM is a project which collects some pre-trained Transformer-based models.
☆12Updated last year
Related projects ⓘ
Alternatives and complementary repositories for CyBERTron-LM
- ☆14Updated last year
- DeFacto - Demonstrations and Feedback for improving factual consistency of text summarization☆27Updated last year
- Fault-aware neural code rankers☆25Updated last year
- The code of EMNLP 2019 paper "A Split-and-Recombine Approach for Follow-up Query Analysis"☆17Updated last year
- UNISUMM: Unified Few-shot Summarization with Multi-Task Pre-Training and Prefix-Tuning☆60Updated last year
- Submission archive for the MS MARCO passage ranking leaderboard☆12Updated last year
- Scripts to parse arxiv documents for NLP tasks☆17Updated last year
- Gallery for Industry AI demos☆17Updated last year
- Official code for "Too Brittle To Touch: Comparing the Stability of Quantization and Distillation Towards Developing Lightweight Low-Reso…☆17Updated last year
- [ICLR 2022] Pretraining Text Encoders with Adversarial Mixture of Training Signal Generators☆24Updated last year
- Generative Retrieval Transformer☆29Updated last year
- Search-based-Neural-Structured-Learning-for-Sequential-Question-Answering☆32Updated last year
- ☆22Updated last year
- We release the UICaption dataset. The dataset consists of UI images (icons and screenshots) and associated text descriptions. This datase…☆35Updated last year
- ☆17Updated 4 years ago
- ☆12Updated 3 years ago
- Boosting Natural Language Generation from Instructions with Meta-Learning☆10Updated last year
- ☆36Updated 3 months ago
- Ongoing research training transformer language models at scale, including: BERT & GPT-2☆18Updated last year
- [NeurIPS 2022] code for "K-LITE: Learning Transferable Visual Models with External Knowledge" https://arxiv.org/abs/2204.09222☆51Updated last year
- ☆49Updated last year
- Towards Semantics-Enhanced Pre-Training: Can Lexicon Definitions Help Learning Sentence Meanings? (AAAI 2021)☆9Updated 3 years ago
- ☆14Updated 3 years ago
- T5Patches is a set of tools for fast and targeted editing of generative language models built with T5X.☆12Updated 5 months ago
- ☆25Updated last year
- AutoMoE: Neural Architecture Search for Efficient Sparsely Activated Transformers☆42Updated 2 years ago
- Resources related to EMNLP 2021 paper "FAME: Feature-Based Adversarial Meta-Embeddings for Robust Input Representations"☆13Updated 2 years ago
- ☆26Updated last year
- BANG is a new pretraining model to Bridge the gap between Autoregressive (AR) and Non-autoregressive (NAR) Generation. AR and NAR generat…☆28Updated 2 years ago
- Code for the paper "Code-Mixing on Sesame Street: Dawn of the Adversarial Polyglots" (NAACL-HLT 2021)☆10Updated 2 years ago