nblt / F-SAM
[CVPR 2024] Friendly Sharpness-Aware Minimization
☆27Updated 3 months ago
Alternatives and similar repositories for F-SAM:
Users that are interested in F-SAM are comparing it to the libraries listed below
- ☆35Updated 2 years ago
- [ICLR 2022] "Anti-Oversmoothing in Deep Vision Transformers via the Fourier Domain Analysis: From Theory to Practice" by Peihao Wang, Wen…☆79Updated last year
- Variance Covariance Regularization☆14Updated last year
- The official repo for CVPR2023 highlight paper "Gradient Norm Aware Minimization Seeks First-Order Flatness and Improves Generalization".☆82Updated last year
- [ICLR 2024] Improving Convergence and Generalization Using Parameter Symmetries☆29Updated 8 months ago
- ☆57Updated 2 years ago
- The offical implement of ImbSAM (Imbalanced-SAM)☆23Updated 11 months ago
- Transformers trained on Tiny ImageNet☆52Updated 2 years ago
- ☆11Updated 2 years ago
- ☆16Updated 2 years ago
- Sharpness-Aware Minimization Leads to Low-Rank Features [NeurIPS 2023]☆27Updated last year
- Code for 'Multi-level Logit Distillation' (CVPR2023)☆57Updated 4 months ago
- ☆62Updated last year
- Compressible Dynamics in Deep Overparameterized Low-Rank Learning & Adaptation (ICML'24 Oral)☆14Updated 6 months ago
- Implementation of ASAM: Adaptive Sharpness-Aware Minimization for Scale-Invariant Learning of Deep Neural Networks, ICML 2021.☆141Updated 3 years ago
- Metrics for "Beyond neural scaling laws: beating power law scaling via data pruning " (NeurIPS 2022 Outstanding Paper Award)☆55Updated last year
- This resposity maintains a collection of important papers on knowledge distillation (awesome-knowledge-distillation)).☆74Updated 2 months ago
- [NeurIPS 2022] Make Sharpness-Aware Minimization Stronger: A Sparsified Perturbation Approach -- Official Implementation☆44Updated last year
- [NeurIPS 2024] AlphaPruning: Using Heavy-Tailed Self Regularization Theory for Improved Layer-wise Pruning of Large Language Models☆21Updated last month
- Code for the paper "Efficient Dataset Distillation using Random Feature Approximation"☆37Updated last year
- SLTrain: a sparse plus low-rank approach for parameter and memory efficient pretraining (NeurIPS 2024)☆30Updated 3 months ago
- Switch EMA: A Free Lunch for Better Flatness and Sharpness☆26Updated last year
- [Preprint] Why is the State of Neural Network Pruning so Confusing? On the Fairness, Comparison Setup, and Trainability in Network Prunin…☆40Updated 2 years ago
- PyTorch implementation of "From Sparse to Soft Mixtures of Experts"☆50Updated last year
- Official PyTorch(MMCV) implementation of “Adversarial AutoMixup” (ICLR 2024 spotlight)☆65Updated 3 months ago
- A generic code base for neural network pruning, especially for pruning at initialization.☆30Updated 2 years ago
- ☆29Updated last year
- [ICLR'23] Trainability Preserving Neural Pruning (PyTorch)☆32Updated last year
- A repository for DenseSSMs☆86Updated 10 months ago
- PyTorch repository for ICLR 2022 paper (GSAM) which improves generalization (e.g. +3.8% top-1 accuracy on ImageNet with ViT-B/32)☆139Updated 2 years ago