Olivia-fsm / DoGE
Codebase for ICML submission "DOGE: Domain Reweighting with Generalization Estimation"
β15Updated last year
Alternatives and similar repositories for DoGE:
Users that are interested in DoGE are comparing it to the libraries listed below
- [ICLR'25 Spotlight] Min-K%++: Improved baseline for detecting pre-training data of LLMsβ35Updated 2 weeks ago
- Skill-It! A Data-Driven Skills Framework for Understanding and Training Language Modelsβ43Updated last year
- AI Logging for Interpretability and Explainabilityπ¬β105Updated 8 months ago
- A curated list of awesome resources dedicated to Scaling Laws for LLMsβ70Updated last year
- β46Updated last year
- Official repository for MATES: Model-Aware Data Selection for Efficient Pretraining with Data Influence Models [NeurIPS 2024]β59Updated 3 months ago
- Code associated with Tuning Language Models by Proxy (Liu et al., 2024)β105Updated 11 months ago
- Official repository for paper "Weak-to-Strong Extrapolation Expedites Alignment"β72Updated 8 months ago
- β89Updated last year
- Learning adapter weights from task descriptionsβ16Updated last year
- Fast and Robust Early-Exiting Framework for Autoregressive Language Models with Synchronized Parallel Decoding (EMNLP 2023 Long)β56Updated 5 months ago
- [ACL'24 Oral] Analysing The Impact of Sequence Composition on Language Model Pre-Trainingβ19Updated 6 months ago
- β45Updated 6 months ago
- Long Context Extension and Generalization in LLMsβ50Updated 5 months ago
- β37Updated last week
- LoFiT: Localized Fine-tuning on LLM Representationsβ33Updated last month
- [NeurIPS'23] Aging with GRACE: Lifelong Model Editing with Discrete Key-Value Adaptorsβ72Updated 2 months ago
- β65Updated 2 years ago
- A Kernel-Based View of Language Model Fine-Tuning https://arxiv.org/abs/2210.05643β74Updated last year
- β47Updated 10 months ago
- Repo accompanying our paper "Do Llamas Work in English? On the Latent Language of Multilingual Transformers".β68Updated 11 months ago
- Official github repo for the paper "Compression Represents Intelligence Linearly" [COLM 2024]β130Updated 5 months ago
- Function Vectors in Large Language Models (ICLR 2024)β140Updated 4 months ago
- The accompanying code for "Transformer Feed-Forward Layers Are Key-Value Memories". Mor Geva, Roei Schuster, Jonathan Berant, and Omer Leβ¦β89Updated 3 years ago
- [ICLR'25] Data and code for our paper "Why Does the Effective Context Length of LLMs Fall Short?"β70Updated 3 months ago
- β22Updated 3 weeks ago
- Repository for "Propagating Knowledge Updates to LMs Through Distillation" (NeurIPS 2023).β25Updated 6 months ago
- Test-time-training on nearest neighbors for large language modelsβ37Updated 10 months ago
- GSM-Plus: Data, Code, and Evaluation for Enhancing Robust Mathematical Reasoning in Math Word Problems.β54Updated 7 months ago