LOG-postech / SasshaLinks
[ICML 2025] Official Pytorch code for "SASSHA: Sharpness-aware Adaptive Second-order Optimization With Stable Hessian Approximation"
β19Updated 3 months ago
Alternatives and similar repositories for Sassha
Users that are interested in Sassha are comparing it to the libraries listed below
Sorting:
- [UAI 2025] Official code for reproducing paper "Critical Influence of Overparameterization on Sharpness-aware Minimization"β19Updated 6 months ago
- π¨ Malet (Machine Learning Experiment Tool) is a tool for efficient machine learning experiment execution, logging, analysis, and plot maβ¦β17Updated 6 months ago
- β17Updated 2 weeks ago
- PyHessian is a Pytorch library for second-order based analysis and training of Neural Networksβ758Updated 4 months ago
- β28Updated 9 months ago
- β25Updated last year
- β17Updated 2 years ago
- [ICLR 2024] Dynamic Sparse Training with Structured Sparsityβ21Updated last year
- tiny-imagenet dataset downloader & reader using tensorflow_datasets (tfds) apiβ20Updated 6 years ago
- Efficient PyTorch Hessian eigendecomposition tools!β382Updated last year
- PyTorch implementation of Mixer-nano (#parameters is 0.67M, originally Mixer-S/16 has 18M) with 90.83 % acc. on CIFAR-10. Training from sβ¦β36Updated 4 years ago
- PyTorch implementation for Vision Transformer[Dosovitskiy, A.(ICLR'21)] modified to obtain over 90% accuracy FROM SCRATCH on CIFAR-10 witβ¦β201Updated last year
- Approximating neural network loss landscapes in low-dimensional parameter subspaces for PyTorchβ348Updated last year
- Lightweight torch implementation of rigl, a sparse-to-sparse optimizer.β60Updated 4 years ago
- A curated list of papers of interesting empirical study and insight on deep learning. Continually updating...β378Updated last week
- This repository contains a Pytorch implementation of the paper "The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks"β¦β337Updated 2 years ago
- Comparison of method "Pruning at initialization prior to training" (Synflow/SNIP/GraSP) in PyTorchβ17Updated last year
- β195Updated 4 years ago
- [KDD 2023] code for "Test accuracy vs. generalization gap: model selection in NLP without accessing training or testing data" https://arxβ¦β12Updated 3 years ago
- Pretrained models on CIFAR10/100 in PyTorchβ374Updated 6 months ago
- [NeurIPS 2021] code for "Taxonomizing local versus global structure in neural network loss landscapes" https://arxiv.org/abs/2107.11228β19Updated 3 years ago
- β72Updated 11 months ago
- Pytorch implementation of the paper "SNIP: Single-shot Network Pruning based on Connection Sensitivity" by Lee et al.β111Updated 6 years ago
- Source code of "Task arithmetic in the tangent space: Improved editing of pre-trained models".β107Updated 2 years ago
- β35Updated 2 years ago
- β227Updated last year
- Train CIFAR10 to 94% accuracy in a few minutes/seconds. Based on https://github.com/davidcpage/cifar10-fastβ22Updated 2 years ago
- EPβ18Updated 4 years ago
- β20Updated last year
- Neural Tangent Kernel Papersβ119Updated 10 months ago