krafton-ai / mini-batch-clLinks
☆11Updated last year
Alternatives and similar repositories for mini-batch-cl
Users that are interested in mini-batch-cl are comparing it to the libraries listed below
Sorting:
- ☆16Updated 2 years ago
- Self-Contrastive Learning: Single-viewed Supervised Contrastive Framework using Sub-network (AAAI 2023)☆20Updated last year
- This is the repository for "Model Merging by Uncertainty-Based Gradient Matching", ICLR 2024.☆27Updated last year
- ☆15Updated 2 months ago
- About Code for the paper "NASH: A Simple Unified Framework of Structured Pruning for Accelerating Encoder-Decoder Language Models" (EMNLP…☆16Updated last year
- Code for "Surgical Fine-Tuning Improves Adaptation to Distribution Shifts" published at ICLR 2023☆28Updated last year
- ☆67Updated 3 years ago
- ☆14Updated 3 years ago
- ☆21Updated 2 years ago
- [NeurIPS 2023] Official repository for "Distilling Out-of-Distribution Robustness from Vision-Language Foundation Models"☆12Updated 11 months ago
- Code for NeurIPS'23 paper "A Bayesian Approach To Analysing Training Data Attribution In Deep Learning"☆17Updated last year
- Code for T-MARS data filtering☆35Updated last year
- Revisiting Efficient Training Algorithms For Transformer-based Language Models (NeurIPS 2023)☆80Updated last year
- Model Stock: All we need is just a few fine-tuned models☆116Updated 8 months ago
- Official code and dataset for our NAACL 2024 paper: DialogCC: An Automated Pipeline for Creating High-Quality Multi-modal Dialogue Datase…☆12Updated 11 months ago
- The official repository for our paper "The Dual Form of Neural Networks Revisited: Connecting Test Time Predictions to Training Patterns …☆16Updated last year
- ☆29Updated 2 years ago
- Domain Adaptation and Adapters☆16Updated 2 years ago
- ☆20Updated 2 years ago
- Official repository of "LiNeS: Post-training Layer Scaling Prevents Forgetting and Enhances Model Merging"☆26Updated 7 months ago
- ☆29Updated 3 months ago
- ☆45Updated 2 years ago
- Introducing Filtered Direct Preference Optimization (fDPO) that enhances language model alignment with human preferences by discarding lo…☆15Updated 6 months ago
- ☆34Updated last week
- Why Do We Need Weight Decay in Modern Deep Learning? [NeurIPS 2024]☆66Updated 8 months ago
- ☆30Updated 10 months ago
- DiWA: Diverse Weight Averaging for Out-of-Distribution Generalization☆31Updated 2 years ago
- ☆128Updated 2 years ago
- Repository for research works and resources related to model reprogramming <https://arxiv.org/abs/2202.10629>☆61Updated last year
- ☆22Updated last year