lutnn / blink-mmLinks
☆15Updated last year
Alternatives and similar repositories for blink-mm
Users that are interested in blink-mm are comparing it to the libraries listed below
Sorting:
- SpInfer: Leveraging Low-Level Sparsity for Efficient Large Language Model Inference on GPUs☆48Updated 3 months ago
- ☆59Updated last year
- ☆148Updated 11 months ago
- ☆18Updated 4 years ago
- PIM-DL: Expanding the Applicability of Commodity DRAM-PIMs for Deep Learning via Algorithm-System Co-Optimization☆31Updated last year
- ☆19Updated 9 months ago
- ☆34Updated last year
- The official implementation of the DAC 2024 paper GQA-LUT☆18Updated 6 months ago
- ☆75Updated 5 months ago
- A curated list for Efficient Large Language Models☆11Updated last year
- PyTorch extension for emulating FP8 data formats on standard FP32 Xeon/GPU hardware.☆110Updated 6 months ago
- ☆38Updated 2 years ago
- [DAC'25] Official implement of "HybriMoE: Hybrid CPU-GPU Scheduling and Cache Management for Efficient MoE Inference"☆54Updated 2 weeks ago
- A GPU-optimized system for efficient long-context LLMs decoding with low-bit KV cache.☆49Updated 2 weeks ago
- ☆51Updated 11 months ago
- Quantized Attention on GPU☆44Updated 7 months ago
- ArkVale: Efficient Generative LLM Inference with Recallable Key-Value Eviction (NIPS'24)☆40Updated 6 months ago
- ☆60Updated 2 months ago
- GPU operators for sparse tensor operations☆33Updated last year
- ☆31Updated last year
- [ICLR 2025] TidalDecode: A Fast and Accurate LLM Decoding with Position Persistent Sparse Attention☆39Updated 2 months ago
- A Vectorized N:M Format for Unleashing the Power of Sparse Tensor Cores☆51Updated last year
- LLM Inference analyzer for different hardware platforms☆74Updated last month
- Magicube is a high-performance library for quantized sparse matrix operations (SpMM and SDDMM) of deep learning on Tensor Cores.☆89Updated 2 years ago
- play gemm with tvm☆91Updated last year
- PipeInfer: Accelerating LLM Inference using Asynchronous Pipelined Speculation☆30Updated 7 months ago
- TileFlow is a performance analysis tool based on Timeloop for fusion dataflows☆61Updated last year
- LLM Inference with Microscaling Format☆23Updated 7 months ago
- DeeperGEMM: crazy optimized version☆69Updated last month
- Decoding Attention is specially optimized for MHA, MQA, GQA and MLA using CUDA core for the decoding stage of LLM inference.☆38Updated 2 weeks ago