ayulockin / debugNNwithWandBLinks
Concepts Explored in/with Pytorch
☆19Updated 10 months ago
Alternatives and similar repositories for debugNNwithWandB
Users that are interested in debugNNwithWandB are comparing it to the libraries listed below
Sorting:
- Collection of snippets for PyTorch users☆25Updated 3 years ago
- Official code for the Stochastic Polyak step-size optimizer☆139Updated last year
- Implémentation of the article **Deep Learning CUDA Memory Usage and Pytorch optimization tricks**☆43Updated 5 years ago
- Module 0 - Fundamentals☆103Updated 9 months ago
- Toy implementations of some popular ML optimizers using Python/JAX☆44Updated 4 years ago
- Simply Numpy implementation of the FAVOR+ attention mechanism, https://teddykoker.com/2020/11/performers/☆38Updated 4 years ago
- Easy-to-use AdaHessian optimizer (PyTorch)☆79Updated 4 years ago
- Implementations and checkpoints for ResNet, Wide ResNet, ResNeXt, ResNet-D, and ResNeSt in JAX (Flax).☆112Updated 3 years ago
- JAX implementation of Learning to learn by gradient descent by gradient descent☆27Updated 8 months ago
- 11-785 Introduction to Deep Learning (IDeeL) website with logistics and select course materials☆43Updated 2 weeks ago
- This is a port of Mistral-7B model in JAX☆32Updated 11 months ago
- Personal solutions to the Triton Puzzles☆19Updated 11 months ago
- A custom PyTorch layer that is capable of implementing extremely wide and sparse linear layers efficiently☆50Updated last year
- Implementation of the Kalman Filtering Attention proposed in "Kalman Filtering Attention for User Behavior Modeling in CTR Prediction"☆58Updated last year
- reproduces experiments from "Grounding inductive biases in natural images: invariance stems from variations in data"☆17Updated 9 months ago
- A minimalistic boiler plate code for training pytorch models☆13Updated last year
- Context Manager to profile the forward and backward times of PyTorch's nn.Module☆83Updated last year
- Materials for "Transformers from the Ground Up" at PyData Jeddah on August 5, 2021☆20Updated 3 years ago
- A set of of fundamental operations and deep learning models using JAX☆12Updated 4 years ago
- ☆36Updated 2 years ago
- The official repository for our paper "The Devil is in the Detail: Simple Tricks Improve Systematic Generalization of Transformers". We s…☆67Updated 2 years ago
- Codes accompanying the paper "LaProp: a Better Way to Combine Momentum with Adaptive Gradient"☆29Updated 4 years ago
- Graph Convolutional Networks in JAX☆32Updated 4 years ago
- This repository hosts the code to port NumPy model weights of BiT-ResNets to TensorFlow SavedModel format.☆14Updated 3 years ago
- ☆68Updated last year
- This repository implements and evaluates convolutional networks on the Möbius strip as toy model instantiations of Coordinate Independent…☆72Updated last year
- Pytorch implementation of a simple way to enable (Stochastic) Frame Averaging for any network☆50Updated 11 months ago
- A simple Transformer where the softmax has been replaced with normalization☆20Updated 4 years ago
- Multiplicative Normalizing Flows in PyTorch.☆24Updated last month
- Generate bird's-eye views of conference proceedings.☆24Updated 6 months ago