JosselinSomervilleRoberts / BERT-Multitask-learning
Multitask-learning of a BERT backbone. Allows to easily train a BERT model with state-of-the-art method such as PCGrad, Gradient Vaccine, PALs, Scheduling, Class imbalance handling and many optimizations
☆18Updated last year
Alternatives and similar repositories for BERT-Multitask-learning:
Users that are interested in BERT-Multitask-learning are comparing it to the libraries listed below
- Fine-tuning LLM with LoRA (Low-Rank Adaptation) from scratch (Oct 2023)☆19Updated last year
- A simple recipe for training and inferencing Transformer architecture for Multi-Task Learning on custom datasets. You can find two approa…☆95Updated 2 years ago
- This repository contains a custom implementation of the BERT model, fine-tuned for specific tasks, along with an implementation of Low Ra…☆70Updated last year
- Scripts for text classification with llama and bert☆14Updated last month
- ☆70Updated 6 months ago
- Text Diffusion Model with Encoder-Decoder Transformers for Sequence-to-Sequence Generation [NAACL 2024]☆93Updated last year
- Minimalist BERT implementation assignment for CS11-711☆81Updated 2 years ago
- [Neurips2023] Source code for Lift Yourself Up: Retrieval-augmented Text Generation with Self Memory☆59Updated last year
- A Simple but Powerful SOTA NER Model | Official Code For Label Supervised LLaMA Finetuning☆154Updated last year
- ☆61Updated 2 years ago
- Tool for converting LLMs from uni-directional to bi-directional by removing causal mask for tasks like classification and sentence embedd…☆57Updated 3 months ago
- Data Augmentation for Intent Classification with Off-the-Shelf Large Language Models is a ServiceNow Research project☆30Updated last year
- domain adaptation in NLP☆53Updated 3 years ago
- A extension of Transformers library to include T5ForSequenceClassification class.☆38Updated last year
- Interpreting Language Models with Contrastive Explanations (EMNLP 2022 Best Paper Honorable Mention)☆62Updated 2 years ago
- Vocabulary Trimming (VT) is a model compression technique, which reduces a multilingual LM vocabulary to a target language by deleting ir…☆35Updated 5 months ago
- Hierarchical Attention Transformers (HAT)☆51Updated last year
- Code for "From Pretraining Data to Language Models to Downstream Tasks: Tracking the Trails of Political Biases Leading to Unfair NLP Mod…☆35Updated last year
- The paper list of multilingual pre-trained models (Continual Updated).☆20Updated 9 months ago
- This repository lists papers, codes, and datasets in Biomedical Text Summarisation based on PLM☆23Updated 2 years ago
- CS 224N Winter 2023 Default Final Project: Multitask BERT☆25Updated 2 years ago
- ☆18Updated 3 years ago
- Leveraging ChatGPT for Text Data Augmentation☆38Updated 6 months ago
- [NeurIPS 2023 Main Track] This is the repository for the paper titled "Don’t Stop Pretraining? Make Prompt-based Fine-tuning Powerful Lea…☆73Updated last year
- Code for ACL paper "Zero-Shot Text Classification via Self-Supervised Tuning"☆27Updated last year
- This is the code for the ICLR 2023 paper "Leveraging Large Language Models for Multiple Choice Question Answering."☆39Updated 2 years ago
- The codebase for the paper: A Closer Look at How Fine-tuning Changes BERT☆22Updated 2 years ago
- This is the oficial repository for "Parameter-Efficient Multi-task Tuning via Attentional Mixtures of Soft Prompts" (EMNLP 2022)☆100Updated 2 years ago
- ☆13Updated 5 months ago
- [ACL 2024] Self-Training with Direct Preference Optimization Improves Chain-of-Thought Reasoning☆41Updated 8 months ago