lernapparat / torchhacksLinks
Hacks for PyTorch
☆19Updated 2 years ago
Alternatives and similar repositories for torchhacks
Users that are interested in torchhacks are comparing it to the libraries listed below
Sorting:
- CUDA implementation of autoregressive linear attention, with all the latest research findings☆44Updated 2 years ago
- ☆29Updated 2 years ago
- Experiment of using Tangent to autodiff triton☆79Updated last year
- A place to store reusable transformer components of my own creation or found on the interwebs☆56Updated last week
- Demo of the unit_scaling library, showing how a model can be easily adapted to train in FP8.☆45Updated 11 months ago
- ☆21Updated 3 months ago
- Triton Implementation of HyperAttention Algorithm☆48Updated last year
- An open source implementation of CLIP.☆32Updated 2 years ago
- ☆12Updated last month
- Make triton easier☆46Updated last year
- ImageNet-12k subset of ImageNet-21k (fall11)☆21Updated 2 years ago
- Context Manager to profile the forward and backward times of PyTorch's nn.Module☆83Updated last year
- Contains my experiments with the `big_vision` repo to train ViTs on ImageNet-1k.☆22Updated 2 years ago
- ☆21Updated last month
- Prototype routines for GPU quantization written using PyTorch.☆21Updated 4 months ago
- FlexAttention w/ FlashAttention3 Support☆26Updated 8 months ago
- Memory-Efficient CUDA kernels for training ConvNets with PyTorch.☆41Updated 4 months ago
- Utilities for PyTorch distributed☆24Updated 3 months ago
- A dashboard for exploring timm learning rate schedulers☆19Updated 7 months ago
- An implementation of the Llama architecture, to instruct and delight☆21Updated 3 weeks ago
- Torch Distributed Experimental☆116Updated 10 months ago
- Implementation of a Transformer using ReLA (Rectified Linear Attention) from https://arxiv.org/abs/2104.07012☆49Updated 3 years ago
- JAX implementation of Learning to learn by gradient descent by gradient descent☆27Updated 8 months ago
- Benchmarking PyTorch 2.0 different models☆21Updated 2 years ago
- Customized matrix multiplication kernels☆56Updated 3 years ago
- A collection of optimizers, some arcane others well known, for Flax.☆29Updated 3 years ago
- Transformer with Mu-Parameterization, implemented in Jax/Flax. Supports FSDP on TPU pods.☆30Updated 2 weeks ago
- Load any clip model with a standardized interface☆21Updated last year
- Implementation of the Remixer Block from the Remixer paper, in Pytorch☆36Updated 3 years ago
- FID computation in Jax/Flax.☆27Updated 11 months ago