Rose-STL-Lab / Teleportation-Optimization
[ICLR 2024] Improving Convergence and Generalization Using Parameter Symmetries
☆29Updated 8 months ago
Alternatives and similar repositories for Teleportation-Optimization:
Users that are interested in Teleportation-Optimization are comparing it to the libraries listed below
- Metrics for "Beyond neural scaling laws: beating power law scaling via data pruning " (NeurIPS 2022 Outstanding Paper Award)☆55Updated last year
- Code release for Deep Incubation (https://arxiv.org/abs/2212.04129)☆90Updated last year
- [Preprint] Why is the State of Neural Network Pruning so Confusing? On the Fairness, Comparison Setup, and Trainability in Network Prunin…☆40Updated 2 years ago
- ☆42Updated last year
- This repository is the implementation of the paper Training Free Pretrained Model Merging (CVPR2024).☆27Updated 11 months ago
- Repository containing code for blockwise SSL training☆28Updated 4 months ago
- [ICLR 2022] "Anti-Oversmoothing in Deep Vision Transformers via the Fourier Domain Analysis: From Theory to Practice" by Peihao Wang, Wen…☆79Updated last year
- Sharpness-Aware Minimization Leads to Low-Rank Features [NeurIPS 2023]☆27Updated last year
- This is a PyTorch implementation of the paperViP A Differentially Private Foundation Model for Computer Vision☆36Updated last year
- ☆21Updated 2 years ago
- Code for paper "Parameter Efficient Multi-task Model Fusion with Partial Linearization"☆18Updated 5 months ago
- Code for ICML2023 paper, DDGR: Continual Learning with Deep Diffusion-based Generative Replay.☆33Updated last year
- The code of the paper "Minimizing the Accumulated Trajectory Error to Improve Dataset Distillation" (CVPR2023)☆40Updated last year
- Data distillation benchmark☆52Updated this week
- [NeurIPS 2022] Make Sharpness-Aware Minimization Stronger: A Sparsified Perturbation Approach -- Official Implementation☆44Updated last year
- ☆42Updated 2 years ago
- Elucidated Dataset Condensation (NeurIPS 2024)☆19Updated 4 months ago
- [ICML2023] Revisiting Data-Free Knowledge Distillation with Poisoned Teachers☆23Updated 7 months ago
- Switch EMA: A Free Lunch for Better Flatness and Sharpness☆26Updated last year
- (NeurIPS 2023 spotlight) Large-scale Dataset Distillation/Condensation, 50 IPC (Images Per Class) achieves the highest 60.8% on original …☆126Updated 3 months ago
- A generic code base for neural network pruning, especially for pruning at initialization.☆30Updated 2 years ago
- Official implementation for paper "Knowledge Diffusion for Distillation", NeurIPS 2023☆80Updated last year
- [CVPR 2024] Friendly Sharpness-Aware Minimization☆27Updated 3 months ago
- ICLR 2024, Towards Lossless Dataset Distillation via Difficulty-Aligned Trajectory Matching☆99Updated 8 months ago
- Official Code for NeurIPS 2022 Paper: How Mask Matters: Towards Theoretical Understandings of Masked Autoencoders☆57Updated last year
- ☆54Updated last year
- [CVPR2024] Efficient Dataset Distillation via Minimax Diffusion☆90Updated 10 months ago
- Official code for "pi-Tuning: Transferring Multimodal Foundation Models with Optimal Multi-task Interpolation", ICML 2023.☆32Updated last year
- Repo for the paper "Extrapolating from a Single Image to a Thousand Classes using Distillation"☆36Updated 7 months ago
- Denoising Masked Autoencoders Help Robust Classification.☆60Updated last year