andbloch / eth-dl-cheat-sheetLinks
Cheat sheet for the "Deep Learning" course at ETH Zürich
☆20Updated 5 years ago
Alternatives and similar repositories for eth-dl-cheat-sheet
Users that are interested in eth-dl-cheat-sheet are comparing it to the libraries listed below
Sorting:
- Code accompanying our paper "Feature Learning in Infinite-Width Neural Networks" (https://arxiv.org/abs/2011.14522)☆62Updated 4 years ago
- PyTorch implementation of "Grokking: Generalization Beyond Overfitting on Small Algorithmic Datasets"☆37Updated 3 years ago
- Repository for my Big Data Optimization course☆34Updated 4 years ago
- ☆63Updated 3 years ago
- Automatic Differentiation for Gradient Boosted Decision Trees.☆13Updated 3 years ago
- unofficial re-implementation of "Grokking: Generalization Beyond Overfitting on Small Algorithmic Datasets"☆78Updated 2 years ago
- Repository for my convex optimization course.☆54Updated 4 years ago
- ☆140Updated 2 weeks ago
- ☆49Updated 4 years ago
- [TMLR 2022] Curvature access through the generalized Gauss-Newton's low-rank structure: Eigenvalues, eigenvectors, directional derivative…☆17Updated last year
- ☆10Updated 2 years ago
- This repository contains a Jax implementation of conformal training corresponding to the ICLR'22 paper "learning optimal conformal classi…☆130Updated 2 years ago
- ☆31Updated 7 months ago
- Transformers with doubly stochastic attention☆46Updated 2 years ago
- LaTeX source code for the slides☆23Updated 3 years ago
- Laplace Redux -- Effortless Bayesian Deep Learning☆42Updated 3 weeks ago
- ☆68Updated 6 months ago
- Omnigrok: Grokking Beyond Algorithmic Data☆58Updated 2 years ago
- Replicating and dissecting the git-re-basin project in one-click-replication Colabs☆36Updated 2 years ago
- ☆24Updated 3 years ago
- Compositional Capabilities of Autoregressive Transformers: A Study on Synthetic, Interpretable Tasks☆10Updated last year
- Code for Accelerated Linearized Laplace Approximation for Bayesian Deep Learning (ELLA, NeurIPS 22')☆16Updated 2 years ago
- Notebooks for managing NeurIPS 2014 and analysing the NeurIPS experiment.☆11Updated last year
- Gradient-based constrained optimization for JAX☆32Updated 2 years ago
- Deep Networks Grok All the Time and Here is Why☆37Updated last year
- General tips to drive your research at Mila☆19Updated last year
- Testing Language Models for Memorization of Tabular Datasets.☆33Updated 4 months ago
- Why Do We Need Weight Decay in Modern Deep Learning? [NeurIPS 2024]☆66Updated 9 months ago
- Codebase for Mechanistic Mode Connectivity☆14Updated last year
- Public Implementation of Volatility Based Kernels and Moving Average Means for Accurate Forecasting with Gaussian Processes☆49Updated 2 years ago