Ankush7890 / ssfinetuning
A package for fine tuning of pretrained NLP transformers using Semi Supervised Learning
☆15Updated 3 years ago
Alternatives and similar repositories for ssfinetuning:
Users that are interested in ssfinetuning are comparing it to the libraries listed below
- ☆14Updated 5 months ago
- ☆29Updated 2 years ago
- Adding new tasks to T0 without catastrophic forgetting☆33Updated 2 years ago
- Starbucks: Improved Training for 2D Matryoshka Embeddings☆18Updated last month
- ☆21Updated 3 years ago
- Implementation of COCO-LM, Correcting and Contrasting Text Sequences for Language Model Pretraining, in Pytorch☆45Updated 4 years ago
- Combining encoder-based language models☆11Updated 3 years ago
- Hugging Face RoBERTa with Flash Attention 2☆22Updated last year
- Code and pre-trained models for "ReasonBert: Pre-trained to Reason with Distant Supervision", EMNLP'2021☆29Updated 2 years ago
- Task Compass: Scaling Multi-task Pre-training with Task Prefix (EMNLP 2022: Findings) (stay tuned & more will be updated)☆22Updated 2 years ago
- Few-shot Learning with Auxiliary Data☆27Updated last year
- Repo for ICML23 "Why do Nearest Neighbor Language Models Work?"☆56Updated 2 years ago
- Embedding Recycling for Language models☆38Updated last year
- A Structured Span Selector (NAACL 2022). A structured span selector with a WCFG for span selection tasks (coreference resolution, semanti…☆21Updated 2 years ago
- An Empirical Study On Contrastive Search And Contrastive Decoding For Open-ended Text Generation☆26Updated 9 months ago
- CSS-LM: Contrastive Semi-supervised Fine-tuning of Pre-trained Language Models☆13Updated last year
- Unifew: Unified Fewshot Learning Model☆18Updated 3 years ago
- ☆46Updated 2 years ago
- codes and pre-trained models of paper "Segatron: Segment-aware Transformer for Language Modeling and Understanding"☆18Updated 2 years ago
- [ICLR 2022] Pretraining Text Encoders with Adversarial Mixture of Training Signal Generators☆24Updated last year
- No Parameters Left Behind: Sensitivity Guided Adaptive Learning Rate for Training Large Transformer Models (ICLR 2022)☆30Updated 3 years ago
- Transformers at any scale☆41Updated last year
- ☆20Updated 2 years ago
- A Benchmark for Robust, Multi-evidence, Multi-answer Question Answering☆16Updated 2 years ago
- ☆12Updated last year
- This repository contains the code for paper Prompting ELECTRA Few-Shot Learning with Discriminative Pre-Trained Models.☆47Updated 2 years ago
- ☆54Updated 2 years ago
- [EMNLP 2022] Language Model Pre-Training with Sparse Latent Typing☆14Updated 2 years ago
- [ACL 2023]: Training Trajectories of Language Models Across Scales https://arxiv.org/pdf/2212.09803.pdf☆23Updated last year
- This repository contains some of the code used in the paper "Training Language Models with Langauge Feedback at Scale"☆27Updated last year