lucidrains / triton-transformerView external linksLinks
Implementation of a Transformer, but completely in Triton
☆279Apr 5, 2022Updated 3 years ago
Alternatives and similar repositories for triton-transformer
Users that are interested in triton-transformer are comparing it to the libraries listed below
Sorting:
- GPT, but made only out of MLPs☆89May 25, 2021Updated 4 years ago
- Implementation of Token Shift GPT - An autoregressive model that solely relies on shifting the sequence space for mixing☆49Jan 27, 2022Updated 4 years ago
- Implementation of fused cosine similarity attention in the same style as Flash Attention☆220Feb 13, 2023Updated 3 years ago
- Contrastive Language-Image Pretraining☆143Sep 6, 2022Updated 3 years ago
- Implementation of TableFormer, Robust Transformer Modeling for Table-Text Encoding, in Pytorch☆39Mar 29, 2022Updated 3 years ago
- A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.☆596Aug 12, 2025Updated 6 months ago
- Kernl lets you run PyTorch transformer models several times faster on GPU with a single line of code, and is designed to be easily hackab…☆1,586Jan 28, 2026Updated 2 weeks ago
- ☆51Jan 28, 2024Updated 2 years ago
- Source-to-Source Debuggable Derivatives in Pure Python☆15Jan 23, 2024Updated 2 years ago
- Implementation of Mega, the Single-head Attention with Multi-headed EMA architecture that currently holds SOTA on Long Range Arena☆207Aug 26, 2023Updated 2 years ago
- GPTQ inference Triton kernel☆321May 18, 2023Updated 2 years ago
- Cataloging released Triton kernels.☆294Sep 9, 2025Updated 5 months ago
- Implementation of Flash Attention in Jax☆225Mar 1, 2024Updated last year
- jax-triton contains integrations between JAX and OpenAI Triton☆439Feb 9, 2026Updated last week
- Transformers components but in Triton☆34May 9, 2025Updated 9 months ago
- [NeurIPS 2023] Sparse Modular Activation for Efficient Sequence Modeling☆40Dec 2, 2023Updated 2 years ago
- An open source implementation of CLIP.☆33Nov 7, 2022Updated 3 years ago
- Source code for "N-ary Constituent Tree Parsing with Recursive Semi-Markov Model" published at ACL 2021☆10May 27, 2021Updated 4 years ago
- My attempts at applying Soundstream design on learned tokenization of text and then applying hierarchical attention to text generation☆90Oct 11, 2024Updated last year
- Type annotations and dynamic checking for a tensor's shape, dtype, names, etc.☆1,471May 2, 2025Updated 9 months ago
- ☆29Oct 3, 2022Updated 3 years ago
- ☆19Dec 4, 2025Updated 2 months ago
- ☆104Nov 7, 2024Updated last year
- A python library for highly configurable transformers - easing model architecture search and experimentation.☆49Nov 30, 2021Updated 4 years ago
- FlexAttention w/ FlashAttention3 Support☆27Oct 5, 2024Updated last year
- A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit and 4-bit floating point (FP8 and FP4) precision on H…☆3,163Updated this week
- Fine-Tuning Pre-trained Transformers into Decaying Fast Weights☆19Oct 9, 2022Updated 3 years ago
- [EMNLP 2023] Official implementation of the algorithm ETSC: Exact Toeplitz-to-SSM Conversion our EMNLP 2023 paper - Accelerating Toeplitz…☆14Oct 17, 2023Updated 2 years ago
- Development repository for the Triton language and compiler☆18,429Updated this week
- [EMNLP 2022] Official implementation of Transnormer in our EMNLP 2022 paper - The Devil in Linear Transformer☆64Jul 30, 2023Updated 2 years ago
- Large Context Attention☆766Oct 13, 2025Updated 4 months ago
- Implementation of π-GAN, for 3d-aware image synthesis, in Pytorch☆124Feb 22, 2021Updated 4 years ago
- Implementation of Hierarchical Transformer Memory (HTM) for Pytorch☆76Sep 15, 2021Updated 4 years ago
- A collection of memory efficient attention operators implemented in the Triton language.☆288Jun 5, 2024Updated last year
- Explorations into the recently proposed Taylor Series Linear Attention☆100Aug 18, 2024Updated last year
- Pytorch library for fast transformer implementations☆1,761Mar 23, 2023Updated 2 years ago
- ☆52Jun 10, 2024Updated last year
- A Python-level JIT compiler designed to make unmodified PyTorch programs faster.☆1,074Apr 17, 2024Updated last year
- ☆21Mar 15, 2023Updated 2 years ago