HPMLL / SpInfer_EuroSys25Links
☆20Updated 4 months ago
Alternatives and similar repositories for SpInfer_EuroSys25
Users that are interested in SpInfer_EuroSys25 are comparing it to the libraries listed below
Sorting:
- ☆33Updated last year
- A GPU-optimized system for efficient long-context LLMs decoding with low-bit KV cache.☆56Updated last week
- Magicube is a high-performance library for quantized sparse matrix operations (SpMM and SDDMM) of deep learning on Tensor Cores.☆89Updated 2 years ago
- A Row Decomposition-based Approach for Sparse Matrix Multiplication on GPUs☆22Updated last year
- ☆150Updated last year
- ☆117Updated last week
- ☆49Updated last year
- ☆89Updated 2 months ago
- ☆16Updated 2 years ago
- ☆109Updated 8 months ago
- ☆42Updated last year
- Flash-LLM: Enabling Cost-Effective and Highly-Efficient Large Generative Model Inference with Unstructured Sparsity☆216Updated last year
- ☆128Updated 7 months ago
- ☆21Updated 4 months ago
- ☆32Updated 2 years ago
- A lightweight design for computation-communication overlap.☆155Updated last month
- A Easy-to-understand TensorOp Matmul Tutorial☆369Updated 10 months ago
- Examples of CUDA implementations by Cutlass CuTe☆214Updated last month
- Github mirror of trition-lang/triton repo.☆48Updated this week
- A baseline repository of Auto-Parallelism in Training Neural Networks☆144Updated 3 years ago
- ArkVale: Efficient Generative LLM Inference with Recallable Key-Value Eviction (NIPS'24)☆42Updated 7 months ago
- ☆18Updated last year
- SpInfer: Leveraging Low-Level Sparsity for Efficient Large Language Model Inference on GPUs☆50Updated 4 months ago
- LLM serving cluster simulator☆108Updated last year
- ☆23Updated last year
- ☆13Updated last year
- Compiler for Dynamic Neural Networks☆46Updated last year
- nnScaler: Compiling DNN models for Parallel Training☆114Updated this week
- Source code of the SC '23 paper: "DASP: Specific Dense Matrix Multiply-Accumulate Units Accelerated General Sparse Matrix-Vector Multipli…☆27Updated last year
- ☆127Updated 2 months ago