google-research / wide-sparse-nets
☆19Updated 3 years ago
Alternatives and similar repositories for wide-sparse-nets:
Users that are interested in wide-sparse-nets are comparing it to the libraries listed below
- ☆25Updated 4 years ago
- ☆36Updated 3 years ago
- Implementation of Kronecker Attention in Pytorch☆18Updated 4 years ago
- ☆17Updated 2 years ago
- Code base for SRSGD.☆28Updated 4 years ago
- Successfully training approximations to full-rank matrices for efficiency in deep learning.☆16Updated 4 years ago
- ☆22Updated 6 years ago
- We investigated corruption robustness across different architectures including Convolutional Neural Networks, Vision Transformers, and th…☆15Updated 3 years ago
- Implementation for ACProp ( Momentum centering and asynchronous update for adaptive gradient methdos, NeurIPS 2021)☆15Updated 3 years ago
- Gradient Starvation: A Learning Proclivity in Neural Networks☆61Updated 4 years ago
- Code accompanying the NeurIPS 2020 paper: WoodFisher (Singh & Alistarh, 2020)☆48Updated 3 years ago
- Codes for the paper "Deep Neural Networks with Multi-Branch Architectures Are Less Non-Convex"☆20Updated 4 years ago
- Code for the CVPR 2021 paper: Understanding Failures of Deep Networks via Robust Feature Extraction☆35Updated 2 years ago
- Code release to reproduce ASHA experiments from "Random Search and Reproducibility for NAS."☆22Updated 5 years ago
- SGD with large step sizes learns sparse features [ICML 2023]☆32Updated last year
- Self-Distillation with weighted ground-truth targets; ResNet and Kernel Ridge Regression☆17Updated 3 years ago
- Factorized Neural Layers☆27Updated last year
- An adaptive training algorithm for residual network☆15Updated 4 years ago
- ☆23Updated 6 years ago
- ☆35Updated last year
- Code accompanying our paper "Finding trainable sparse networks through Neural Tangent Transfer" to be published at ICML-2020.☆13Updated 4 years ago
- Experiments from "The Generalization-Stability Tradeoff in Neural Network Pruning": https://arxiv.org/abs/1906.03728.☆14Updated 4 years ago
- Code for "Training Neural Networks with Fixed Sparse Masks" (NeurIPS 2021).☆58Updated 3 years ago
- ☆57Updated 2 years ago
- [ICML2022] Training Your Sparse Neural Network Better with Any Mask. Ajay Jaiswal, Haoyu Ma, Tianlong Chen, ying Ding, and Zhangyang Wang☆27Updated 2 years ago
- ☆11Updated 2 years ago
- Public Codebase for Rethinking Parameter Counting: Effective Dimensionality Revisited☆37Updated 2 years ago
- [JMLR] TRADES + random smoothing for certifiable robustness☆14Updated 4 years ago
- Delta Orthogonal Initialization for PyTorch☆18Updated 6 years ago
- ICML 2020, Estimating Generalization under Distribution Shifts via Domain-Invariant Representations☆22Updated 4 years ago