JoelNiklaus / loss_landscape
Code for visualizing the loss landscape of neural nets
☆10Updated 4 years ago
Alternatives and similar repositories for loss_landscape:
Users that are interested in loss_landscape are comparing it to the libraries listed below
- source code for paper "Riemannian Preconditioned LoRA for Fine-Tuning Foundation Models"☆24Updated 10 months ago
- Deep Learning & Information Bottleneck☆60Updated last year
- [ICLR 2024] Dynamic Sparse Training with Structured Sparsity☆17Updated last year
- Official implementation of ORCA proposed in the paper "Cross-Modal Fine-Tuning: Align then Refine"☆71Updated last year
- Official implementation for Equivariant Architectures for Learning in Deep Weight Spaces [ICML 2023]☆89Updated last year
- Recycling diverse models☆44Updated 2 years ago
- Code for the paper "Efficient Dataset Distillation using Random Feature Approximation"☆37Updated 2 years ago
- Robust Principles: Architectural Design Principles for Adversarially Robust CNNs☆22Updated last year
- SLTrain: a sparse plus low-rank approach for parameter and memory efficient pretraining (NeurIPS 2024)☆30Updated 5 months ago
- ☆11Updated 2 years ago
- Towards Understanding the Mixture-of-Experts Layer in Deep Learning☆27Updated last year
- A simple Jax implementation of influence functions.☆16Updated last year
- Official implementation for Sparse MetA-Tuning (SMAT)☆16Updated 9 months ago
- [ICLR'24] "DeepZero: Scaling up Zeroth-Order Optimization for Deep Model Training" by Aochuan Chen*, Yimeng Zhang*, Jinghan Jia, James Di…☆57Updated 6 months ago
- Source code of "What can linearized neural networks actually say about generalization?☆20Updated 3 years ago
- Code for "The Expressive Power of Low-Rank Adaptation".☆20Updated last year
- Data-free knowledge distillation using Gaussian noise (NeurIPS paper)☆15Updated 2 years ago
- ☆17Updated 2 years ago
- Decoupled Kullback-Leibler Divergence Loss (DKL), NeurIPS 2024 / Generalized Kullback-Leibler Divergence Loss (GKL)☆44Updated 2 weeks ago
- [NeurIPS‘2021] "MEST: Accurate and Fast Memory-Economic Sparse Training Framework on the Edge", Geng Yuan, Xiaolong Ma, Yanzhi Wang et al…☆18Updated 3 years ago
- Code to reproduce experiments from 'Does Knowledge Distillation Really Work' a paper which appeared in the 2021 NeurIPS proceedings.☆33Updated last year
- This repository is the official implementation of Generalized Data Weighting via Class-level Gradient Manipulation (NeurIPS 2021)(http://…☆24Updated 2 years ago
- ModelDiff: A Framework for Comparing Learning Algorithms☆56Updated last year
- Official Implementation of the CVPR'23 paper 'Regularization of polynomial networks for image recognition'.☆9Updated last year
- ☆37Updated 8 months ago
- Official PyTorch implementation of "Meta-prediction Model for Distillation-Aware NAS on Unseen Datasets" (ICLR 2023 notable top 25%)☆24Updated last year
- ☆13Updated 2 years ago
- ☆23Updated 2 years ago
- Deep Networks Grok All the Time and Here is Why☆34Updated 11 months ago
- [NeurIPS 2022] Make Sharpness-Aware Minimization Stronger: A Sparsified Perturbation Approach -- Official Implementation☆44Updated last year