lernapparat / torchhacksLinks
Hacks for PyTorch
☆19Updated 2 years ago
Alternatives and similar repositories for torchhacks
Users that are interested in torchhacks are comparing it to the libraries listed below
Sorting:
- Experiment of using Tangent to autodiff triton☆80Updated last year
- CUDA implementation of autoregressive linear attention, with all the latest research findings☆46Updated 2 years ago
- A place to store reusable transformer components of my own creation or found on the interwebs☆62Updated 2 weeks ago
- Demo of the unit_scaling library, showing how a model can be easily adapted to train in FP8.☆46Updated last year
- Customized matrix multiplication kernels☆57Updated 3 years ago
- Torch Distributed Experimental☆117Updated last year
- ☆29Updated 3 years ago
- Make triton easier☆49Updated last year
- Context Manager to profile the forward and backward times of PyTorch's nn.Module☆83Updated 2 years ago
- Prototype routines for GPU quantization written using PyTorch.☆21Updated 4 months ago
- ☆21Updated 9 months ago
- MaskedTensors for PyTorch☆38Updated 3 years ago
- Triton Implementation of HyperAttention Algorithm☆48Updated last year
- TorchFix - a linter for PyTorch-using code with autofix support☆151Updated 3 months ago
- An open source implementation of CLIP.☆33Updated 3 years ago
- PyTorch centric eager mode debugger☆48Updated 11 months ago
- a lightweight transformer library for PyTorch☆72Updated 4 years ago
- FID computation in Jax/Flax.☆29Updated last year
- A dashboard for exploring timm learning rate schedulers☆19Updated last year
- ☆19Updated 3 years ago
- FlexAttention w/ FlashAttention3 Support☆27Updated last year
- pytest plugin for a better developer experience when working with the PyTorch test suite☆44Updated 3 years ago
- Quantize transformers to any learned arbitrary 4-bit numeric format☆49Updated 4 months ago
- A block oriented training approach for inference time optimization.☆33Updated last year
- No-GIL Python environment featuring NVIDIA Deep Learning libraries.☆69Updated 7 months ago
- A case study of efficient training of large language models using commodity hardware.☆68Updated 3 years ago
- ImageNet-12k subset of ImageNet-21k (fall11)☆21Updated 2 years ago
- Experimental scripts for researching data adaptive learning rate scheduling.☆22Updated 2 years ago
- Little article showing how to load pytorch's models with linear memory consumption☆34Updated 3 years ago
- Implementation of a Transformer using ReLA (Rectified Linear Attention) from https://arxiv.org/abs/2104.07012☆49Updated 3 years ago