g-benton / loss-surface-simplexesLinks
β100Updated 3 years ago
Alternatives and similar repositories for loss-surface-simplexes
Users that are interested in loss-surface-simplexes are comparing it to the libraries listed below
Sorting:
- Codebase for Learning Invariances in Neural Networksβ96Updated 3 years ago
- π§ Pytorch code for the Fromage optimiser.β128Updated last year
- A Machine Learning workflow for Slurm.β151Updated 4 years ago
- Hessian spectral density estimation in TF and Jaxβ124Updated 5 years ago
- β133Updated 4 years ago
- A library for evaluating representations.β77Updated 3 years ago
- PyTorch-SSO: Scalable Second-Order methods in PyTorchβ147Updated 2 years ago
- Bayesianize: A Bayesian neural network wrapper in pytorchβ89Updated last year
- β245Updated 2 years ago
- β54Updated last year
- Experiments for Meta-Learning Symmetries by Reparameterizationβ57Updated 4 years ago
- β153Updated 5 years ago
- Easy-to-use AdaHessian optimizer (PyTorch)β79Updated 4 years ago
- Implicit MLE: Backpropagating Through Discrete Exponential Family Distributionsβ258Updated last year
- {KFAC,EKFAC,Diagonal,Implicit} Fisher Matrices and finite width NTKs in PyTorchβ215Updated 3 weeks ago
- Differentiable Sorting Networksβ117Updated 2 years ago
- Stochastic Automatic Differentiation library for PyTorch.β208Updated last year
- π© Pytorch and Jax code for the Madam optimiser.β52Updated 4 years ago
- Cyclical Stochastic Gradient MCMC for Bayesian Deep Learningβ92Updated 4 years ago
- Convolutional Neural Tangent Kernelβ112Updated 5 years ago
- Code for NeurIPS 2019 paper: "Tensor Programs I: Wide Feedforward or Recurrent Neural Networks of Any Architecture are Gaussian Processesβ¦β247Updated 5 years ago
- Riemannian Convex Potential Mapsβ67Updated 2 years ago
- Code for the paper: "Tensor Programs II: Neural Tangent Kernel for Any Architecture"β105Updated 5 years ago
- This repository contains the results for the paper: "Descending through a Crowded Valley - Benchmarking Deep Learning Optimizers"β184Updated 4 years ago
- Hypergradient descentβ149Updated last year
- The original code for the paper "How to train your MAML" along with a replication of the original "Model Agnostic Meta Learning" (MAML) pβ¦β41Updated 4 years ago
- Official code for the Stochastic Polyak step-size optimizerβ139Updated last year
- Supporting code for the paper "Dangers of Bayesian Model Averaging under Covariate Shift"β33Updated 2 years ago
- β64Updated last year
- codebase for "A Theory of the Inductive Bias and Generalization of Kernel Regression and Wide Neural Networks"β49Updated 2 years ago