zhisbug / Cavs
Cavs: An Efficient Runtime System for Dynamic Neural Networks
☆13Updated 4 years ago
Alternatives and similar repositories for Cavs:
Users that are interested in Cavs are comparing it to the libraries listed below
- An Attention Superoptimizer☆21Updated last week
- ☆23Updated 2 months ago
- Tacker: Tensor-CUDA Core Kernel Fusion for Improving the GPU Utilization while Ensuring QoS☆18Updated 3 years ago
- An IR for efficiently simulating distributed ML computation.☆25Updated last year
- ☆44Updated last year
- CUDA Templates for Linear Algebra Subroutines☆12Updated this week
- Supplemental materials for The ASPLOS 2025 / EuroSys 2025 Contest on Intra-Operator Parallelism for Distributed Deep Learning☆21Updated last month
- ☆21Updated last year
- (NeurIPS 2022) Automatically finding good model-parallel strategies, especially for complex models and clusters.☆37Updated 2 years ago
- TileFusion is a highly efficient kernel template library designed to elevate the level of abstraction in CUDA C for processing tiles.☆43Updated this week
- ☆16Updated 2 years ago
- Benchmark PyTorch Custom Operators☆13Updated last year
- ☆73Updated 2 years ago
- ☆11Updated 3 years ago
- ☆8Updated last year
- An extention of TVMScript to write simple and high performance GPU kernels with tensorcore.☆51Updated 6 months ago
- ☆24Updated last year
- Mille Crepe Bench: layer-wise performance analysis for deep learning frameworks.☆17Updated 5 years ago
- FlexFlow Serve: Low-Latency, High-Performance LLM Serving☆17Updated this week
- DietCode Code Release☆61Updated 2 years ago
- A source-to-source compiler for optimizing CUDA dynamic parallelism by aggregating launches☆14Updated 5 years ago
- ☆48Updated 7 months ago
- GVProf: A Value Profiler for GPU-based Clusters☆48Updated 10 months ago
- An external memory allocator example for PyTorch.☆14Updated 3 years ago
- Benchmark for matrix multiplications between dense and block sparse (BSR) matrix in TVM, blocksparse (Gray et al.) and cuSparse.☆25Updated 4 years ago
- Graphiler is a compiler stack built on top of DGL and TorchScript which compiles GNNs defined using user-defined functions (UDFs) into ef…☆61Updated 2 years ago
- FTPipe and related pipeline model parallelism research.☆41Updated last year
- Thunder Research Group's Collective Communication Library☆31Updated 9 months ago
- PET: Optimizing Tensor Programs with Partially Equivalent Transformations and Automated Corrections☆117Updated 2 years ago
- This is the implementation for paper: AdaTune: Adaptive Tensor Program CompilationMade Efficient (NeurIPS 2020).☆13Updated 3 years ago