jiazhihao / attention_superoptimizer
An Attention Superoptimizer
☆20Updated 8 months ago
Alternatives and similar repositories for attention_superoptimizer:
Users that are interested in attention_superoptimizer are comparing it to the libraries listed below
- Tacker: Tensor-CUDA Core Kernel Fusion for Improving the GPU Utilization while Ensuring QoS☆18Updated 3 years ago
- Cavs: An Efficient Runtime System for Dynamic Neural Networks☆13Updated 4 years ago
- ☆36Updated this week
- An external memory allocator example for PyTorch.☆14Updated 3 years ago
- PyTorch compilation tutorial covering TorchScript, torch.fx, and Slapo☆19Updated last year
- Official resporitory for "IPDPS' 24 QSync: Quantization-Minimized Synchronous Distributed Training Across Hybrid Devices".☆19Updated 10 months ago
- ☆24Updated last year
- Supplemental materials for The ASPLOS 2025 / EuroSys 2025 Contest on Intra-Operator Parallelism for Distributed Deep Learning☆21Updated last month
- (NeurIPS 2022) Automatically finding good model-parallel strategies, especially for complex models and clusters.☆37Updated 2 years ago
- A source-to-source compiler for optimizing CUDA dynamic parallelism by aggregating launches☆14Updated 5 years ago
- ☆8Updated last year
- ☆16Updated 2 years ago
- Artifact for "Apparate: Rethinking Early Exits to Tame Latency-Throughput Tensions in ML Serving" [SOSP '24]☆22Updated last month
- ☆12Updated 2 years ago
- ☆19Updated 3 months ago
- ☆44Updated last year
- FlexFlow Serve: Low-Latency, High-Performance LLM Serving☆15Updated this week
- ThrillerFlow is a Dataflow Analysis and Codegen Framework written in Rust.☆14Updated last month
- ☆13Updated 2 years ago
- A memory profiler for NVIDIA GPUs to explore memory inefficiencies in GPU-accelerated applications.☆22Updated 3 months ago
- Stateful LLM Serving☆44Updated 5 months ago
- SOTA Learning-augmented Systems☆34Updated 2 years ago
- ☆11Updated 3 years ago
- A GPU-accelerated DNN inference serving system that supports instant kernel preemption and biased concurrent execution in GPU scheduling.☆40Updated 2 years ago
- DISB is a new DNN inference serving benchmark with diverse workloads and models, as well as real-world traces.☆53Updated 4 months ago
- Benchmark for matrix multiplications between dense and block sparse (BSR) matrix in TVM, blocksparse (Gray et al.) and cuSparse.☆25Updated 4 years ago
- ☆23Updated last month
- ☆48Updated 7 months ago
- Artifact of ASPLOS'23 paper entitled: GRACE: A Scalable Graph-Based Approach to Accelerating Recommendation Model Inference☆17Updated last year
- GPTQ inference TVM kernel☆38Updated 8 months ago