exists-forall / striped_attention
☆38Updated last year
Alternatives and similar repositories for striped_attention:
Users that are interested in striped_attention are comparing it to the libraries listed below
- PyTorch bindings for CUTLASS grouped GEMM.☆64Updated 3 months ago
- ☆100Updated 5 months ago
- Odysseus: Playground of LLM Sequence Parallelism☆64Updated 8 months ago
- ☆59Updated 2 weeks ago
- Boosting 4-bit inference kernels with 2:4 Sparsity☆64Updated 5 months ago
- ☆72Updated 3 years ago
- Triton-based implementation of Sparse Mixture of Experts.☆196Updated 2 months ago
- Official repository for LightSeq: Sequence Level Parallelism for Distributed Training of Long Context Transformers☆205Updated 6 months ago
- Tritonbench is a collection of PyTorch custom operators with example inputs to measure their performance.☆89Updated this week
- Python package for rematerialization-aware gradient checkpointing☆24Updated last year
- ☆67Updated 3 months ago
- TileFusion is a highly efficient kernel template library designed to elevate the level of abstraction in CUDA C for processing tiles.☆56Updated this week
- ☆22Updated last year
- PyTorch bindings for CUTLASS grouped GEMM.☆94Updated last month
- GEAR: An Efficient KV Cache Compression Recipefor Near-Lossless Generative Inference of LLM☆157Updated 7 months ago
- Ouroboros: Speculative Decoding with Large Model Enhanced Drafting (EMNLP 2024 main)☆84Updated 4 months ago
- (NeurIPS 2022) Automatically finding good model-parallel strategies, especially for complex models and clusters.☆37Updated 2 years ago
- A minimal implementation of vllm.☆33Updated 6 months ago
- High-speed GEMV kernels, at most 2.7x speedup compared to pytorch baseline.☆97Updated 7 months ago
- Transformers components but in Triton☆31Updated 3 months ago
- extensible collectives library in triton☆83Updated 4 months ago
- ☆61Updated 3 weeks ago
- ☆44Updated last month
- Benchmark code for the "Online normalizer calculation for softmax" paper☆66Updated 6 years ago
- ☆81Updated 5 months ago
- GPTQ inference TVM kernel☆38Updated 9 months ago
- 16-fold memory access reduction with nearly no loss☆76Updated 3 months ago
- The source code of our work "Prepacking: A Simple Method for Fast Prefilling and Increased Throughput in Large Language Models"☆59Updated 4 months ago
- Sirius, an efficient correction mechanism, which significantly boosts Contextual Sparsity models on reasoning tasks while maintaining its…☆21Updated 5 months ago
- [NeurIPS 2024] The official implementation of "Kangaroo: Lossless Self-Speculative Decoding for Accelerating LLMs via Double Early Exitin…☆48Updated 7 months ago