Ankush7890 / ssfinetuning
A package for fine tuning of pretrained NLP transformers using Semi Supervised Learning
☆15Updated 3 years ago
Alternatives and similar repositories for ssfinetuning:
Users that are interested in ssfinetuning are comparing it to the libraries listed below
- Embedding Recycling for Language models☆38Updated last year
- Adding new tasks to T0 without catastrophic forgetting☆33Updated 2 years ago
- ☆46Updated 3 years ago
- ☆29Updated 2 years ago
- ☆14Updated 6 months ago
- ☆21Updated 3 years ago
- [ACL 2023] Few-shot Reranking for Multi-hop QA via Language Model Prompting☆27Updated last year
- Starbucks: Improved Training for 2D Matryoshka Embeddings☆19Updated 2 months ago
- Implementation of COCO-LM, Correcting and Contrasting Text Sequences for Language Model Pretraining, in Pytorch☆45Updated 4 years ago
- Transformers at any scale☆41Updated last year
- Combining encoder-based language models☆11Updated 3 years ago
- This repository contains the code for paper Prompting ELECTRA Few-Shot Learning with Discriminative Pre-Trained Models.☆47Updated 2 years ago
- No Parameters Left Behind: Sensitivity Guided Adaptive Learning Rate for Training Large Transformer Models (ICLR 2022)☆30Updated 3 years ago
- HyPe: Better Pre-trained Language Model Fine-tuning with Hidden Representation Perturbation [ACL 2023]☆14Updated last year
- codes and pre-trained models of paper "Segatron: Segment-aware Transformer for Language Modeling and Understanding"☆18Updated 2 years ago
- ☆54Updated 2 years ago
- Ranking of fine-tuned HF models as base models.☆35Updated last year
- Task Compass: Scaling Multi-task Pre-training with Task Prefix (EMNLP 2022: Findings) (stay tuned & more will be updated)☆22Updated 2 years ago
- ☆11Updated 4 months ago
- Hugging Face RoBERTa with Flash Attention 2☆22Updated last year
- ☆13Updated 2 years ago
- Few-shot Learning with Auxiliary Data☆27Updated last year
- Implementation of the model: "Reka Core, Flash, and Edge: A Series of Powerful Multimodal Language Models" in PyTorch☆30Updated last week
- Interpretable unified language safety checking with large language models☆30Updated last year
- Code and pre-trained models for "ReasonBert: Pre-trained to Reason with Distant Supervision", EMNLP'2021☆29Updated 2 years ago
- [ICLR 2022] Pretraining Text Encoders with Adversarial Mixture of Training Signal Generators☆24Updated last year
- official repo of AAAI2024 paper Mitigating the Impact of False Negatives in Dense Retrieval with Contrastive Confidence Regularization☆13Updated last year
- [ACL 2023]: Training Trajectories of Language Models Across Scales https://arxiv.org/pdf/2212.09803.pdf☆23Updated last year
- Tasks for describing differences between text distributions.☆16Updated 8 months ago
- No Parameter Left Behind: How Distillation and Model Size Affect Zero-Shot Retrieval☆29Updated 2 years ago