krafton-ai / mini-batch-cl
☆11Updated last year
Alternatives and similar repositories for mini-batch-cl
Users that are interested in mini-batch-cl are comparing it to the libraries listed below
Sorting:
- ☆16Updated last year
- ☆20Updated 2 years ago
- Self-Contrastive Learning: Single-viewed Supervised Contrastive Framework using Sub-network (AAAI 2023)☆20Updated last year
- Code for "Surgical Fine-Tuning Improves Adaptation to Distribution Shifts" published at ICLR 2023☆28Updated last year
- ☆21Updated 2 years ago
- The git repository of Modular Prompted Chatbot paper☆34Updated last year
- ☆14Updated 3 years ago
- This is the repository for "Model Merging by Uncertainty-Based Gradient Matching", ICLR 2024.☆27Updated last year
- Censored Sampling of Diffusion Models Using 3 Minutes of Human Feedback☆26Updated last year
- ☆34Updated last month
- Domain Adaptation and Adapters☆16Updated 2 years ago
- ☆66Updated 3 years ago
- Repository for research works and resources related to model reprogramming <https://arxiv.org/abs/2202.10629>☆62Updated last year
- ☆11Updated 4 months ago
- Code for the paper "Data Feedback Loops: Model-driven Amplification of Dataset Biases"☆16Updated 2 years ago
- ☆25Updated 7 months ago
- Code for "SAM as an Optimal Relaxation of Bayes", ICLR 2023.☆25Updated last year
- ☆45Updated 2 years ago
- Official code and dataset for our NAACL 2024 paper: DialogCC: An Automated Pipeline for Creating High-Quality Multi-modal Dialogue Datase…☆12Updated 10 months ago
- Code for "Just Train Twice: Improving Group Robustness without Training Group Information"☆71Updated last year
- ☆30Updated 10 months ago
- Why Do We Need Weight Decay in Modern Deep Learning? [NeurIPS 2024]☆66Updated 7 months ago
- DiWA: Diverse Weight Averaging for Out-of-Distribution Generalization☆30Updated 2 years ago
- Code and Dataset release of "Carpe Diem: On the Evaluation of World Knowledge in Lifelong Language Models" (NAACL 2024)☆10Updated 7 months ago
- LISA for ICML 2022☆49Updated 2 years ago
- [NeurIPS 2023] Official repository for "Distilling Out-of-Distribution Robustness from Vision-Language Foundation Models"☆12Updated 11 months ago
- Code for NeurIPS'23 paper "A Bayesian Approach To Analysing Training Data Attribution In Deep Learning"☆17Updated last year
- Official repository of "LiNeS: Post-training Layer Scaling Prevents Forgetting and Enhances Model Merging"☆26Updated 6 months ago
- Switch EMA: A Free Lunch for Better Flatness and Sharpness☆26Updated last year
- Introducing Filtered Direct Preference Optimization (fDPO) that enhances language model alignment with human preferences by discarding lo…☆14Updated 5 months ago