pytorch / maskedtensor
MaskedTensors for PyTorch
☆38Updated 2 years ago
Alternatives and similar repositories for maskedtensor
Users that are interested in maskedtensor are comparing it to the libraries listed below
Sorting:
- CUDA implementation of autoregressive linear attention, with all the latest research findings☆44Updated last year
- ☆29Updated 2 years ago
- Implementation of some personal helper functions for Einops, my most favorite tensor manipulation library ❤️☆54Updated 2 years ago
- Another attempt at a long-context / efficient transformer by me☆38Updated 3 years ago
- ☆33Updated 2 years ago
- Experiment of using Tangent to autodiff triton☆78Updated last year
- ☆31Updated last month
- ☆53Updated 7 months ago
- Blog post☆17Updated last year
- Code for the paper PermuteFormer☆42Updated 3 years ago
- Implementation of LogAvgExp for Pytorch☆36Updated last month
- ☆32Updated 7 months ago
- Very deep VAEs in JAX/Flax☆46Updated 3 years ago
- ☆103Updated 10 months ago
- AdaCat☆49Updated 2 years ago
- Implementation of Gated State Spaces, from the paper "Long Range Language Modeling via Gated State Spaces", in Pytorch☆100Updated 2 years ago
- Implementation of "compositional attention" from MILA, a multi-head attention variant that is reframed as a two-step attention process wi…☆51Updated 3 years ago
- ☆32Updated last year
- GPT, but made only out of MLPs☆88Updated 3 years ago
- Implementation of Hourglass Transformer, in Pytorch, from Google and OpenAI☆89Updated 3 years ago
- A selection of neural network models ported from torchvision for JAX & Flax.☆44Updated 4 years ago
- Official code for "Accelerating Feedforward Computation via Parallel Nonlinear Equation Solving", ICML 2021☆27Updated 3 years ago
- Parallel Associative Scan for Language Models☆18Updated last year
- AdamW optimizer for bfloat16 models in pytorch 🔥.☆32Updated 11 months ago
- Amos optimizer with JEstimator lib.☆82Updated last year
- code for "Semi-Discrete Normalizing Flows through Differentiable Tessellation"☆26Updated 2 years ago
- Demo of the unit_scaling library, showing how a model can be easily adapted to train in FP8.☆45Updated 10 months ago
- Why Do We Need Weight Decay in Modern Deep Learning? [NeurIPS 2024]☆66Updated 7 months ago
- Implementation of a Transformer using ReLA (Rectified Linear Attention) from https://arxiv.org/abs/2104.07012☆50Updated 3 years ago
- Unofficial but Efficient Implementation of "Mamba: Linear-Time Sequence Modeling with Selective State Spaces" in JAX☆83Updated last year