tridao / cutlass_quant
☆16Updated this week
Related projects: ⓘ
- ☆23Updated 9 months ago
- Odysseus: Playground of LLM Sequence Parallelism☆50Updated 3 months ago
- GPTQ inference TVM kernel☆35Updated 4 months ago
- IntLLaMA: A fast and light quantization solution for LLaMA☆18Updated last year
- ☆28Updated 3 months ago
- TensorRT LLM Benchmark Configuration☆10Updated last month
- APPy (Annotated Parallelism for Python) enables users to annotate loops and tensor expressions in Python with compiler directives akin to…☆20Updated 5 months ago
- PyTorch bindings for CUTLASS grouped GEMM.☆41Updated 3 weeks ago
- ☆29Updated 4 months ago
- ☆14Updated last week
- ☆15Updated last month
- ☆11Updated last year
- ☆83Updated 3 weeks ago
- GPU operators for sparse tensor operations☆27Updated 6 months ago
- Inference framework for MoE layers based on TensorRT with Python binding☆41Updated 3 years ago
- Patch convolution to avoid large GPU memory usage of Conv2D☆73Updated 3 months ago
- Official implementation of ICML 2024 paper "ExCP: Extreme LLM Checkpoint Compression via Weight-Momentum Joint Shrinking".☆37Updated 2 months ago
- ☆50Updated 3 months ago
- (NeurIPS 2022) Automatically finding good model-parallel strategies, especially for complex models and clusters.☆33Updated last year
- A parallelism VAE avoids OOM for high resolution image generation☆34Updated 2 months ago
- ☆27Updated 3 weeks ago
- A Suite for Parallel Inference of Diffusion Transformers (DiTs) on multi-GPU Clusters☆25Updated last month
- [ICLR 2024] Jaiswal, A., Gan, Z., Du, X., Zhang, B., Wang, Z., & Yang, Y. Compressing llms: The truth is rarely pure and never simple.☆15Updated 6 months ago
- ☆67Updated last week
- An external memory allocator example for PyTorch.☆13Updated 2 years ago
- CUDA 12.2 HMM demos☆16Updated last month
- [ACL 2024] RelayAttention for Efficient Large Language Model Serving with Long System Prompts☆29Updated 6 months ago
- Pytorch implementation of our paper accepted by ICML 2024 -- CaM: Cache Merging for Memory-efficient LLMs Inference☆21Updated 3 months ago
- Standalone Flash Attention v2 kernel without libtorch dependency☆93Updated last week
- ☆38Updated 9 months ago