declare-lab / della
DELLA-Merging: Reducing Interference in Model Merging through Magnitude-Based Sampling
☆27Updated 4 months ago
Related projects ⓘ
Alternatives and complementary repositories for della
- Data and code for our paper "Why Does the Effective Context Length of LLMs Fall Short?"☆61Updated last week
- ☆62Updated last month
- Official repository for paper "Weak-to-Strong Extrapolation Expedites Alignment"☆67Updated 5 months ago
- ☆29Updated last year
- Code associated with Tuning Language Models by Proxy (Liu et al., 2024)☆96Updated 7 months ago
- An Efficient LLM Fine-Tuning Factory Optimized for MoE PEFT☆40Updated last month
- [NeurIPS2024] Twin-Merging: Dynamic Integration of Modular Expertise in Model Merging☆33Updated 3 weeks ago
- A block pruning framework for LLMs.☆12Updated 4 months ago
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆129Updated last month
- Repo for the EMNLP'24 Paper "Dual-Space Knowledge Distillation for Large Language Models".☆36Updated last week
- [NeurIPS 2024 Spotlight] EMR-Merging: Tuning-Free High-Performance Model Merging☆30Updated 3 weeks ago
- Official PyTorch implementation of DistiLLM: Towards Streamlined Distillation for Large Language Models (ICML 2024)☆134Updated last month
- In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination Mitigation (ICML 2024)☆45Updated 7 months ago
- Code and Data for "Long-context LLMs Struggle with Long In-context Learning"☆91Updated 4 months ago
- The Paper List on Data Contamination for Large Language Models Evaluation.☆74Updated this week
- [ICLR 2024 Spotlight] Code for the paper "Merge, Then Compress: Demystify Efficient SMoE with Hints from Its Routing Policy"☆64Updated 5 months ago
- [NeurIPS'24] Official code for *🎯DART-Math: Difficulty-Aware Rejection Tuning for Mathematical Problem-Solving*☆75Updated last month
- Exploring Model Kinship for Merging Large Language Models☆19Updated last week
- State-of-the-art Parameter-Efficient MoE Fine-tuning Method☆90Updated 2 months ago
- A Survey on the Honesty of Large Language Models☆44Updated last month
- [EMNLP Findings 2024 & ACL 2024 NLRSE Oral] Enhancing Mathematical Reasoning in Language Models with Fine-grained Rewards☆44Updated 6 months ago
- ☆150Updated 9 months ago
- Unofficial Implementation of Evolutionary Model Merging☆33Updated 7 months ago
- ☆53Updated 2 months ago
- ☆27Updated last year
- [NAACL 2024 Outstanding Paper] Source code for the NAACL 2024 paper entitled "R-Tuning: Instructing Large Language Models to Say 'I Don't…☆83Updated 4 months ago
- [EMNLP 2024 (Oral)] Leave No Document Behind: Benchmarking Long-Context LLMs with Extended Multi-Doc QA☆90Updated this week
- [ACL 2024] The official codebase for the paper "Self-Distillation Bridges Distribution Gap in Language Model Fine-tuning".☆97Updated last week
- The official repository of the Omni-MATH benchmark.☆47Updated last week
- Pytorch implementation for "Compressed Context Memory For Online Language Model Interaction" (ICLR'24)☆49Updated 6 months ago