IBM / model-recycling
Ranking of fine-tuned HF models as base models.
☆35Updated last year
Alternatives and similar repositories for model-recycling:
Users that are interested in model-recycling are comparing it to the libraries listed below
- Embedding Recycling for Language models☆38Updated last year
- SWIM-IR is a Synthetic Wikipedia-based Multilingual Information Retrieval training set with 28 million query-passage pairs spanning 33 la…☆47Updated last year
- ☆44Updated 4 months ago
- ☆22Updated 3 years ago
- ☆11Updated 3 months ago
- ☆14Updated 5 months ago
- Few-shot Learning with Auxiliary Data☆27Updated last year
- SMASHED is a toolkit designed to apply transformations to samples in datasets, such as fields extraction, tokenization, prompting, batchi…☆32Updated 10 months ago
- This is the official PyTorch repo for "UNIREX: A Unified Learning Framework for Language Model Rationale Extraction" (ICML 2022).☆24Updated 2 years ago
- This repo contains code for the paper "Psychologically-informed chain-of-thought prompts for metaphor understanding in large language mod…☆14Updated last year
- ☆12Updated 6 months ago
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.☆93Updated 2 years ago
- Source code and data for Like a Good Nearest Neighbor☆28Updated 2 months ago
- ☆46Updated 2 years ago
- Plug-and-play Search Interfaces with Pyserini and Hugging Face☆31Updated last year
- Companion Repo for the Vision Language Modelling YouTube series - https://bit.ly/3PsbsC2 - by Prithivi Da. Open to PRs and collaborations☆14Updated 2 years ago
- Using short models to classify long texts☆21Updated 2 years ago
- Code for paper "Do Language Models Have Beliefs? Methods for Detecting, Updating, and Visualizing Model Beliefs"☆28Updated 2 years ago
- QAmeleon introduces synthetic multilingual QA data using PaLM, a 540B large language model. This dataset was generated by prompt tuning P…☆34Updated last year
- ☆21Updated 3 years ago
- A package for fine tuning of pretrained NLP transformers using Semi Supervised Learning☆15Updated 3 years ago
- Starbucks: Improved Training for 2D Matryoshka Embeddings☆19Updated last month
- CCQA A New Web-Scale Question Answering Dataset for Model Pre-Training☆32Updated 2 years ago
- Repo for ICML23 "Why do Nearest Neighbor Language Models Work?"☆56Updated 2 years ago
- ☆21Updated 2 months ago
- ☆19Updated 2 years ago
- This project develops compact transformer models tailored for clinical text analysis, balancing efficiency and performance for healthcare…☆18Updated last year
- Code for the paper: Saying No is An Art: Contextualized Fallback Responses for Unanswerable Dialogue Queries☆19Updated 3 years ago
- Code and pre-trained models for "ReasonBert: Pre-trained to Reason with Distant Supervision", EMNLP'2021☆29Updated 2 years ago
- Code for equipping pretrained language models (BART, GPT-2, XLNet) with commonsense knowledge for generating implicit knowledge statement…☆16Updated 3 years ago