tanganke / peta
Code for paper "Parameter Efficient Multi-task Model Fusion with Partial Linearization"
☆13Updated last month
Related projects ⓘ
Alternatives and complementary repositories for peta
- Representation Surgery for Multi-Task Model Merging. ICML, 2024.☆28Updated last month
- AdaMerging: Adaptive Model Merging for Multi-Task Learning. ICLR, 2024.☆49Updated 2 weeks ago
- Official repository of "Localizing Task Information for Improved Model Merging and Compression" [ICML 2024]☆34Updated 2 weeks ago
- Official code for "pi-Tuning: Transferring Multimodal Foundation Models with Optimal Multi-task Interpolation", ICML 2023.☆32Updated last year
- [NeurIPS 2024 Spotlight] EMR-Merging: Tuning-Free High-Performance Model Merging☆30Updated 3 weeks ago
- ☆23Updated 7 months ago
- ☆31Updated last year
- source code of (quasi-)Givens Orthogonal Fine Tuning integrated to peft lib☆13Updated last month
- EfficientVLM: Fast and Accurate Vision-Language Models via Knowledge Distillation and Modal-adaptive Pruning (ACL 2023)☆20Updated last year
- ☆44Updated 10 months ago
- Source code of "Task arithmetic in the tangent space: Improved editing of pre-trained models".☆85Updated last year
- This is the official repo for Debiasing Large Visual Language Models, including a Post-Hoc debias method and Visual Debias Decoding strat…☆71Updated 7 months ago
- A curated list of Model Merging methods.☆82Updated last month
- [NeurIPS 2023] Code release for "Going Beyond Linear Mode Connectivity: The Layerwise Linear Feature Connectivity"☆16Updated last year
- This is the repository for "Model Merging by Uncertainty-Based Gradient Matching", ICLR 2024.☆20Updated 5 months ago
- [ICML 2024] Junk DNA Hypothesis: A Task-Centric Angle of LLM Pre-trained Weights through Sparsity; Lu Yin*, Ajay Jaiswal*, Shiwei Liu, So…☆15Updated 5 months ago
- Awesome-Low-Rank-Adaptation☆34Updated 3 weeks ago
- Source code of EMNLP 2022 Findings paper "SparseAdapter: An Easy Approach for Improving the Parameter-Efficiency of Adapters"☆19Updated 7 months ago
- source code for NeurIPS'24 paper "HaloScope: Harnessing Unlabeled LLM Generations for Hallucination Detection"☆18Updated last month
- [ACL 2023] Code for paper “Tailoring Instructions to Student’s Learning Levels Boosts Knowledge Distillation”(https://arxiv.org/abs/2305.…☆37Updated last year
- Dataset pruning for ImageNet and LAION-2B.☆68Updated 4 months ago
- Official Pytorch Implementation of Our Paper Accepted at ICLR 2024-- Dynamic Sparse No Training: Training-Free Fine-tuning for Sparse LLM…☆36Updated 7 months ago
- Codes for Merging Large Language Models☆24Updated 3 months ago
- ☆34Updated 3 months ago
- ☆35Updated 3 months ago
- ☆19Updated last month
- Source code for the TMLR paper "Black-Box Prompt Learning for Pre-trained Language Models"☆55Updated last year
- Code for paper: “What Data Benefits My Classifier?” Enhancing Model Performance and Interpretability through Influence-Based Data Selecti…☆22Updated 5 months ago
- Data-free knowledge distillation using Gaussian noise (NeurIPS paper)☆15Updated last year
- Implementaiton of "DiLM: Distilling Dataset into Language Model for Text-level Dataset Distillation" (accepted by NAACL2024 Findings)".☆14Updated last week