AristotelisPap / Question-Answering-with-BERT-and-Knowledge-DistillationLinks
Fine-tuned BERT on SQuAd 2.0 Dataset. Applied Knowledge Distillation (KD) and fine-tuned DistilBERT (student) using BERT as the teacher model. Reduced the size of the original BERT by 40%.
☆25Updated 4 years ago
Alternatives and similar repositories for Question-Answering-with-BERT-and-Knowledge-Distillation
Users that are interested in Question-Answering-with-BERT-and-Knowledge-Distillation are comparing it to the libraries listed below
Sorting:
- ☆27Updated 6 months ago
- Code associated with the "Data Augmentation using Pre-trained Transformer Models" paper☆52Updated 2 years ago
- Use BERT for Question Answering and finetune train with SQuAD 2.0☆15Updated 5 years ago
- This repository contains the code for paper Prompting ELECTRA Few-Shot Learning with Discriminative Pre-Trained Models.☆48Updated 3 years ago
- pyTorch implementation of Recurrence over BERT (RoBERT) based on this paper https://arxiv.org/abs/1910.10781 and comparison with pyTorch …☆82Updated 2 years ago
- The source code of "Language Models are Few-shot Multilingual Learners" (MRL @ EMNLP 2021)☆52Updated 3 years ago
- A repository for our AAAI-2020 Cross-lingual-NER paper. Code will be updated shortly.☆47Updated 2 years ago
- The PyTorch implementation of ReCoSa(the Relevant Contexts with Self-attention) for dialogue generation using the multi-head attention an…☆22Updated 2 years ago
- ☆42Updated 4 years ago
- Codes for our CCL 2021 paper: Incorporating Commonsense Knowledge into Abstractive Dialogue Summarization via Heterogeneous Graph Network…☆25Updated 3 years ago
- The sources codes of the DR-BERT model and baselines☆38Updated 3 years ago
- ☆33Updated 2 years ago
- Source codes of Neural Quality Estimation with Multiple Hypotheses for Grammatical Error Correction☆43Updated 3 years ago
- Neural Fuzzy Repair (NFR) is a data augmentation pipeline, which integrates fuzzy matches (i.e. similar translations) into neural machine…☆11Updated 10 months ago
- ☆92Updated 3 years ago
- [ACL 2020] Structure-Level Knowledge Distillation For Multilingual Sequence Labeling☆71Updated 2 years ago
- ☆66Updated 3 years ago
- Code associated with the "Data Augmentation using Pre-trained Transformer Models" paper☆134Updated 2 years ago
- [NAACL'22] TaCL: Improving BERT Pre-training with Token-aware Contrastive Learning☆93Updated 3 years ago
- Using BERT For Classifying Documents with Long Texts, check my latest post: https://armandolivares.tech/☆41Updated 5 years ago
- This repo supports various cross-lingual transfer learning & multilingual NLP models.☆92Updated last year
- ☆30Updated 4 years ago
- Official repository for the paper "Question Answering Infused Pre-training of General-Purpose Contextualized Representations" by Robin Ji…☆15Updated 3 years ago
- Prompt-learning methods used BERT4Keras (PET, EFL and NSP-BERT), both for Chinese and English.☆29Updated 2 years ago
- Bert for CoQA based on PyTorch.☆43Updated 2 years ago
- CoSDA-ML: Multi-Lingual Code-Switching Data Augmentation for Zero-Shot Cross-Lingual NLP☆52Updated 3 years ago
- ☆71Updated 3 years ago
- ☆38Updated 2 years ago
- ☆42Updated 3 years ago
- This is a simple implementation of how to leverage a Language Model for a prompt-based learning model☆44Updated 3 years ago