chips-compilers-mlsys-21 / chips-compilers-mlsys-21.github.io
☆11Updated 3 years ago
Related projects: ⓘ
- Tacker: Tensor-CUDA Core Kernel Fusion for Improving the GPU Utilization while Ensuring QoS☆17Updated 2 years ago
- An external memory allocator example for PyTorch.☆13Updated 2 years ago
- ☆14Updated last week
- An Attention Superoptimizer☆19Updated 4 months ago
- PSTensor provides a way to hack the memory management of tensors in TensorFlow and PyTorch by defining your own C++ Tensor Class.☆9Updated 2 years ago
- ☆45Updated last year
- An IR for efficiently simulating distributed ML computation.☆24Updated 8 months ago
- ☆23Updated 6 months ago
- PET: Optimizing Tensor Programs with Partially Equivalent Transformations and Automated Corrections☆112Updated 2 years ago
- Benchmark scripts for TVM☆73Updated 2 years ago
- Benchmark PyTorch Custom Operators☆13Updated last year
- Slides from 2021-12-15 talk, "TVM Developer Bootcamp – Writing Hardware Backends"☆10Updated 2 years ago
- ☆7Updated last year
- A standalone GEMM kernel for fp16 activation and quantized weight, extracted from FasterTransformer☆82Updated 6 months ago
- An optimizing compiler for decision tree ensemble inference.☆15Updated last week
- Fairring (FAIR + Herring) is a plug-in for PyTorch that provides a process group for distributed training that outperforms NCCL at large …☆61Updated 2 years ago
- Chameleon: Adaptive Code Optimization for Expedited Deep Neural Network Compilation☆26Updated 4 years ago
- Official resporitory for "IPDPS' 24 QSync: Quantization-Minimized Synchronous Distributed Training Across Hybrid Devices".☆19Updated 6 months ago
- GPTQ inference TVM kernel☆35Updated 4 months ago
- PyTorch compilation tutorial covering TorchScript, torch.fx, and Slapo☆18Updated last year
- TensorRT LLM Benchmark Configuration☆10Updated last month
- An extention of TVMScript to write simple and high performance GPU kernels with tensorcore.☆49Updated last month
- play gemm with tvm☆81Updated last year
- ☆52Updated this week
- A memory profiler for NVIDIA GPUs to explore memory inefficiencies in GPU-accelerated applications.☆20Updated 3 months ago
- Yet another Polyhedra Compiler for DeepLearning☆19Updated last year
- Inference framework for MoE layers based on TensorRT with Python binding☆41Updated 3 years ago
- TiledCUDA is a highly efficient kernel template library designed to elevate CUDA C’s level of abstraction for processing tiles.☆114Updated last week
- (NeurIPS 2022) Automatically finding good model-parallel strategies, especially for complex models and clusters.☆33Updated last year