Olivia-fsm / DoGELinks
Codebase for ICML submission "DOGE: Domain Reweighting with Generalization Estimation"
☆17Updated last year
Alternatives and similar repositories for DoGE
Users that are interested in DoGE are comparing it to the libraries listed below
Sorting:
- ☆67Updated 3 years ago
- Learning adapter weights from task descriptions☆18Updated last year
- Test-time-training on nearest neighbors for large language models☆41Updated last year
- Implementation of Gradient Information Optimization (GIO) for effective and scalable training data selection☆14Updated last year
- ☆49Updated last year
- Stanford NLP Python library for benchmarking the utility of LLM interpretability methods☆92Updated this week
- Skill-It! A Data-Driven Skills Framework for Understanding and Training Language Models☆46Updated last year
- ☆48Updated last year
- Is In-Context Learning Sufficient for Instruction Following in LLMs? [ICLR 2025]☆30Updated 4 months ago
- Code for "Tracing Knowledge in Language Models Back to the Training Data"☆38Updated 2 years ago
- Official code repo for paper "Great Memory, Shallow Reasoning: Limits of kNN-LMs"☆23Updated last month
- Fast and Robust Early-Exiting Framework for Autoregressive Language Models with Synchronized Parallel Decoding (EMNLP 2023 Long)☆60Updated 8 months ago
- Repository for "Propagating Knowledge Updates to LMs Through Distillation" (NeurIPS 2023).☆25Updated 9 months ago
- Restore safety in fine-tuned language models through task arithmetic☆28Updated last year
- Code associated with Tuning Language Models by Proxy (Liu et al., 2024)☆111Updated last year
- In-context Example Selection with Influences☆15Updated 2 years ago
- Augmenting Statistical Models with Natural Language Parameters☆26Updated 8 months ago
- Exploration of automated dataset selection approaches at large scales.☆42Updated 3 months ago
- [ICML 2024] Junk DNA Hypothesis: A Task-Centric Angle of LLM Pre-trained Weights through Sparsity; Lu Yin*, Ajay Jaiswal*, Shiwei Liu, So…☆16Updated last month
- [NAACL 2025] A Closer Look into Mixture-of-Experts in Large Language Models☆52Updated 3 months ago
- Long Is More for Alignment: A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning [ICML 2024]☆17Updated last year
- Long Context Extension and Generalization in LLMs☆56Updated 8 months ago
- ☆94Updated last year
- ☆23Updated 3 months ago
- AI Logging for Interpretability and Explainability🔬☆119Updated 11 months ago
- Revisiting Efficient Training Algorithms For Transformer-based Language Models (NeurIPS 2023)☆80Updated last year
- EMNLP 2024: Model Editing Harms General Abilities of Large Language Models: Regularization to the Rescue☆35Updated last week
- [NeurIPS'23] Aging with GRACE: Lifelong Model Editing with Discrete Key-Value Adaptors☆77Updated 5 months ago
- Providing the answer to "How to do patching on all available SAEs on GPT-2?". It is an official repository of the implementation of the p…☆11Updated 4 months ago
- [ICLR'25 Spotlight] Min-K%++: Improved baseline for detecting pre-training data of LLMs☆37Updated last week