JoelNiklaus / loss_landscapeLinks
Code for visualizing the loss landscape of neural nets
☆10Updated 4 years ago
Alternatives and similar repositories for loss_landscape
Users that are interested in loss_landscape are comparing it to the libraries listed below
Sorting:
- [ICLR 2024] Dynamic Sparse Training with Structured Sparsity☆18Updated last year
- source code for paper "Riemannian Preconditioned LoRA for Fine-Tuning Foundation Models"☆27Updated last year
- Parameter-Efficient Fine-Tuning of State Space Models (ICML 2025)☆17Updated last month
- Robust Principles: Architectural Design Principles for Adversarially Robust CNNs☆24Updated last year
- Official repo of Progressive Data Expansion: data, code and evaluation☆29Updated last year
- Recycling diverse models☆45Updated 2 years ago
- Official implementation of ORCA proposed in the paper "Cross-Modal Fine-Tuning: Align then Refine"☆71Updated last year
- ☆16Updated 2 years ago
- ☆19Updated last year
- ☆12Updated 3 years ago
- ☆23Updated 2 years ago
- ☆15Updated 3 years ago
- Deep Learning & Information Bottleneck☆61Updated 2 years ago
- Code for "Don't trust your eyes: on the (un)reliability of feature visualizations" (ICML 2024)☆32Updated last year
- ☆15Updated 2 years ago
- Code for experiments on transformers using Markovian data.☆17Updated 7 months ago
- Code and results accompanying our paper titled RLSbench: Domain Adaptation under Relaxed Label Shift☆34Updated last year
- This is the public github for our paper "Transformer with a Mixture of Gaussian Keys"☆28Updated 2 years ago
- Decoupled Kullback-Leibler Divergence Loss (DKL), NeurIPS 2024 / Generalized Kullback-Leibler Divergence Loss (GKL)☆44Updated last month
- Official PyTorch implementation of "Meta-prediction Model for Distillation-Aware NAS on Unseen Datasets" (ICLR 2023 notable top 25%)☆24Updated last year
- ☆23Updated 2 years ago
- ☆13Updated 2 years ago
- SLTrain: a sparse plus low-rank approach for parameter and memory efficient pretraining (NeurIPS 2024)☆32Updated 8 months ago
- ☆37Updated 11 months ago
- Official implementation for Equivariant Architectures for Learning in Deep Weight Spaces [ICML 2023]☆89Updated last year
- Applies ROME and MEMIT on Mamba-S4 models☆14Updated last year
- Deep Networks Grok All the Time and Here is Why☆37Updated last year
- Why Do We Need Weight Decay in Modern Deep Learning? [NeurIPS 2024]☆66Updated 9 months ago
- ModelDiff: A Framework for Comparing Learning Algorithms☆59Updated last year
- [NeurIPS '24] Code repo for the paper entitled "Learning Structured Representations with Hyperbolic Embeddings" at NeurIPS 2024☆15Updated 5 months ago