fKunstner / noise-sgd-adam-sign
☆16Updated 2 years ago
Alternatives and similar repositories for noise-sgd-adam-sign
Users that are interested in noise-sgd-adam-sign are comparing it to the libraries listed below
Sorting:
- A modern look at the relationship between sharpness and generalization [ICML 2023]☆43Updated last year
- SGD with large step sizes learns sparse features [ICML 2023]☆32Updated 2 years ago
- Implementations of orthogonal and semi-orthogonal convolutions in the Fourier domain with applications to adversarial robustness☆44Updated 4 years ago
- Towards Understanding Sharpness-Aware Minimization [ICML 2022]☆35Updated 2 years ago
- ☆17Updated 2 years ago
- ☆13Updated 2 years ago
- Official code for "In Search of Robust Measures of Generalization" (NeurIPS 2020)☆28Updated 4 years ago
- ☆58Updated 2 years ago
- [ICML 2024] SINGD: KFAC-like Structured Inverse-Free Natural Gradient Descent (http://arxiv.org/abs/2312.05705)☆21Updated 6 months ago
- Code for "SAM as an Optimal Relaxation of Bayes", ICLR 2023.☆25Updated last year
- Code base for SRSGD.☆28Updated 5 years ago
- ☆27Updated 2 years ago
- Official code for "Accelerating Feedforward Computation via Parallel Nonlinear Equation Solving", ICML 2021☆27Updated 3 years ago
- ☆25Updated 4 years ago
- ☆41Updated 2 years ago
- Source code of "What can linearized neural networks actually say about generalization?☆20Updated 3 years ago
- Pytorch code for "Improving Self-Supervised Learning by Characterizing Idealized Representations"☆41Updated 2 years ago
- ☆53Updated 9 months ago
- ☆30Updated 4 years ago
- ☆23Updated 2 years ago
- Code for ICLR 2022 Paper, "Controlling Directions Orthogonal to a Classifier"☆35Updated last year
- [NeurIPS'20] Code for the Paper Compositional Visual Generation and Inference with Energy Based Models☆44Updated 2 years ago
- [JMLR] TRADES + random smoothing for certifiable robustness☆14Updated 4 years ago
- Codes for the paper "Optimizing Mode Connectivity via Neuron Alignment" from NeurIPS 2020.☆16Updated 4 years ago
- Training vision models with full-batch gradient descent and regularization☆37Updated 2 years ago
- Code for Accelerated Linearized Laplace Approximation for Bayesian Deep Learning (ELLA, NeurIPS 22')☆16Updated 2 years ago
- Code to reproduce experiments from 'Does Knowledge Distillation Really Work' a paper which appeared in the 2021 NeurIPS proceedings.☆33Updated last year
- Supporing code for the paper "Bayesian Model Selection, the Marginal Likelihood, and Generalization".☆35Updated 2 years ago
- Large-batch Training, Neural Network Optimization☆9Updated 5 years ago
- DiWA: Diverse Weight Averaging for Out-of-Distribution Generalization☆30Updated 2 years ago