K-Wu / pytorch-direct
Code for Large Graph Convolutional Network Training with GPU-Oriented Data Communication Architecture (accepted by PVLDB).The outdated write-up (https://arxiv.org/abs/2101.07956) explains engineering details, but only a portion of the functionality is migrated to this newer PyTorch version 1.8.0nightly (e152ca5).
☆9Updated last year
Alternatives and similar repositories for pytorch-direct:
Users that are interested in pytorch-direct are comparing it to the libraries listed below
- An external memory allocator example for PyTorch.☆14Updated 3 years ago
- A memory profiler for NVIDIA GPUs to explore memory inefficiencies in GPU-accelerated applications.☆25Updated 6 months ago
- Tacker: Tensor-CUDA Core Kernel Fusion for Improving the GPU Utilization while Ensuring QoS☆25Updated 2 months ago
- An Attention Superoptimizer☆21Updated 3 months ago
- GPTQ inference TVM kernel☆38Updated last year
- PSTensor provides a way to hack the memory management of tensors in TensorFlow and PyTorch by defining your own C++ Tensor Class.☆10Updated 3 years ago
- Official resporitory for "IPDPS' 24 QSync: Quantization-Minimized Synchronous Distributed Training Across Hybrid Devices".☆19Updated last year
- TileFusion is an experimental C++ macro kernel template library that elevates the abstraction level in CUDA C for tile processing.☆82Updated this week
- ☆23Updated 5 months ago
- FractalTensor is a programming framework that introduces a novel approach to organizing data in deep neural networks (DNNs) as a list of …☆26Updated 4 months ago
- ☆19Updated 6 months ago
- ☆9Updated last year
- ☆11Updated 4 years ago
- Cavs: An Efficient Runtime System for Dynamic Neural Networks☆14Updated 4 years ago
- Fairring (FAIR + Herring) is a plug-in for PyTorch that provides a process group for distributed training that outperforms NCCL at large …☆65Updated 3 years ago
- ☆38Updated 5 years ago
- CUDA 12.2 HMM demos☆19Updated 8 months ago
- An extention of TVMScript to write simple and high performance GPU kernels with tensorcore.☆50Updated 9 months ago
- ☆23Updated 2 months ago
- ☆50Updated last year
- play gemm with tvm☆90Updated last year
- Benchmark scripts for TVM☆74Updated 3 years ago
- ☆9Updated 2 years ago
- 方便扩展的Cuda算子理解和优化框架,仅用在学习使用☆13Updated 10 months ago
- This is a demo how to write a high performance convolution run on apple silicon☆54Updated 3 years ago
- Standalone Flash Attention v2 kernel without libtorch dependency☆108Updated 7 months ago
- An IR for efficiently simulating distributed ML computation.☆28Updated last year
- Supplemental materials for The ASPLOS 2025 / EuroSys 2025 Contest on Intra-Operator Parallelism for Distributed Deep Learning☆23Updated 4 months ago
- A standalone GEMM kernel for fp16 activation and quantized weight, extracted from FasterTransformer☆91Updated 3 weeks ago
- Mille Crepe Bench: layer-wise performance analysis for deep learning frameworks.☆17Updated 5 years ago