alibaba / BladeDISC
BladeDISC is an end-to-end DynamIc Shape Compiler project for machine learning workloads.
☆847Updated 2 months ago
Alternatives and similar repositories for BladeDISC:
Users that are interested in BladeDISC are comparing it to the libraries listed below
- A flexible and efficient deep neural network (DNN) compiler that generates high-performance executable from a DNN model description.☆978Updated 5 months ago
- A model compilation solution for various hardware☆409Updated last week
- A primitive library for neural network☆1,318Updated 3 months ago
- row-major matmul optimization☆608Updated last year
- MegCC是一个运行时超轻量,高效,移植简单的深度学习模型编译器☆480Updated 4 months ago
- optimized BERT transformer inference on NVIDIA GPU. https://arxiv.org/abs/2210.03052☆469Updated 11 months ago
- GLake: optimizing GPU memory management and IO transmission.☆435Updated 3 months ago
- Easy Parallel Library (EPL) is a general and efficient deep learning framework for distributed model training.☆266Updated last year
- FlagGems is an operator library for large language models implemented in Triton Language.☆445Updated this week
- ☆407Updated this week
- Yinghan's Code Sample☆313Updated 2 years ago
- heterogeneity-aware-lowering-and-optimization☆254Updated last year
- ☆194Updated last year
- TensorRT Plugin Autogen Tool☆369Updated last year
- Development repository for the Triton-Linalg conversion☆176Updated last month
- Dive into Deep Learning Compiler☆647Updated 2 years ago
- The Tensor Algebra SuperOptimizer for Deep Learning☆703Updated 2 years ago
- ☆601Updated 9 months ago
- ppl.cv is a high-performance image processing library of openPPL supporting various platforms.☆500Updated 4 months ago
- A CPU tool for benchmarking the peak of floating points☆527Updated 5 months ago
- Compiler Infrastructure for Neural Networks☆145Updated last year
- Running BERT without Padding☆473Updated 2 years ago
- High performance Cross-platform Inference-engine, you could run Anakin on x86-cpu,arm, nv-gpu, amd-gpu,bitmain and cambricon devices.☆533Updated 2 years ago
- Representation and Reference Lowering of ONNX Models in MLIR Compiler Infrastructure☆825Updated this week
- A simple high performance CUDA GEMM implementation.☆350Updated last year
- A fast communication-overlapping library for tensor/expert parallelism on GPUs.☆397Updated this week
- DeepLearning Framework Performance Profiling Toolkit☆285Updated 2 years ago
- The Torch-MLIR project aims to provide first class support from the PyTorch ecosystem to the MLIR ecosystem.☆1,459Updated this week
- TePDist (TEnsor Program DISTributed) is an HLO-level automatic distributed system for DL models.☆91Updated last year
- ☆319Updated last month