uiuctml / MergeBenchLinks
MergeBench: A Benchmark for Merging Domain-Specialized LLMs
☆17Updated 2 months ago
Alternatives and similar repositories for MergeBench
Users that are interested in MergeBench are comparing it to the libraries listed below
Sorting:
- State-of-the-art paired encoder and decoder models (17M-1B params)☆25Updated this week
- Official code release for "SuperBPE: Space Travel for Language Models"☆61Updated this week
- Simple and scalable tools for data-driven pretraining data selection.☆24Updated last month
- Official code repo for paper "Great Memory, Shallow Reasoning: Limits of kNN-LMs"☆23Updated 2 months ago
- This repository contains the code used for the experiments in the paper "Fine-Tuning Enhances Existing Mechanisms: A Case Study on Entity…☆27Updated last year
- https://footprints.baulab.info☆17Updated 9 months ago
- Code for the paper "Getting the most out of your tokenizer for pre-training and domain adaptation"☆19Updated last year
- ☆38Updated last year
- ☆14Updated last month
- [ACL'24 Oral] Analysing The Impact of Sequence Composition on Language Model Pre-Training☆22Updated 11 months ago
- Reference implementation for Reward-Augmented Decoding: Efficient Controlled Text Generation With a Unidirectional Reward Model☆43Updated last year
- Minimum Bayes Risk Decoding for Hugging Face Transformers☆58Updated last year
- ☆27Updated 5 months ago
- ☆35Updated 2 years ago
- Few-shot Learning with Auxiliary Data☆28Updated last year
- ☆11Updated last year
- This is the official implementation for our ACL 2024 paper: "Causal Estimation of Memorisation Profiles".☆23Updated 3 months ago
- A Kernel-Based View of Language Model Fine-Tuning https://arxiv.org/abs/2210.05643☆76Updated last year
- Landing page for MIB: A Mechanistic Interpretability Benchmark☆16Updated last week
- Code for Zero-Shot Tokenizer Transfer☆133Updated 6 months ago
- ☆51Updated last year
- ☆54Updated 2 years ago
- Language models scale reliably with over-training and on downstream tasks☆97Updated last year
- Adding new tasks to T0 without catastrophic forgetting☆33Updated 2 years ago
- ☆19Updated last year
- Repo for ICML23 "Why do Nearest Neighbor Language Models Work?"☆58Updated 2 years ago
- IntructIR, a novel benchmark specifically designed to evaluate the instruction following ability in information retrieval models. Our foc…☆32Updated last year
- ☆45Updated last year
- Code for NeurIPS 2024 Spotlight: "Scaling Laws and Compute-Optimal Training Beyond Fixed Training Durations"☆77Updated 8 months ago
- Evaluation pipeline for the BabyLM Challenge 2023.☆76Updated last year