onnx / digestaiLinks
Digest AI is a powerful model analysis tool that extracts insights from your models.
☆32Updated 4 months ago
Alternatives and similar repositories for digestai
Users that are interested in digestai are comparing it to the libraries listed below
Sorting:
- AI Tensor Engine for ROCm☆284Updated this week
- Home for OctoML PyTorch Profiler☆114Updated 2 years ago
- Intel® Extension for DeepSpeed* is an extension to DeepSpeed that brings feature support with SYCL kernels on Intel GPU(XPU) device. Note…☆63Updated 3 months ago
- MLPerf™ logging library☆37Updated this week
- Ahead of Time (AOT) Triton Math Library☆78Updated this week
- OpenAI Triton backend for Intel® GPUs☆210Updated this week
- MLIR-based partitioning system☆136Updated this week
- TritonParse: A Compiler Tracer, Visualizer, and Reproducer for Triton Kernels☆155Updated this week
- An experimental CPU backend for Triton (https//github.com/openai/triton)☆45Updated last month
- Machine Learning Agility (MLAgility) benchmark and benchmarking tools☆40Updated 2 months ago
- Benchmarks to capture important workloads.☆31Updated 8 months ago
- Unified compiler/runtime for interfacing with PyTorch Dynamo.☆101Updated last month
- Development repository for the Triton language and compiler☆133Updated this week
- oneCCL Bindings for Pytorch*☆102Updated 2 months ago
- An IR for efficiently simulating distributed ML computation.☆29Updated last year
- No-code CLI designed for accelerating ONNX workflows☆215Updated 4 months ago
- ☆58Updated this week
- TORCH_LOGS parser for PT2☆61Updated 3 weeks ago
- LeetGPU Challenges☆104Updated last week
- AMD RAD's multi-GPU Triton-based framework for seamless multi-GPU programming☆82Updated last week
- ☆28Updated 3 months ago
- A fork of tvm/unity☆14Updated 2 years ago
- ArcticInference: vLLM plugin for high-throughput, low-latency inference☆270Updated this week
- Fast low-bit matmul kernels in Triton☆379Updated 2 weeks ago
- Dev repo for power measurement for the MLPerf™ benchmarks☆24Updated last month
- Issues related to MLPerf™ Inference policies, including rules and suggested changes☆64Updated last month
- Model compression for ONNX☆97Updated 10 months ago
- Notes and artifacts from the ONNX steering committee☆26Updated this week
- A Fusion Code Generator for NVIDIA GPUs (commonly known as "nvFuser")☆357Updated this week
- A lightweight MLIR Python frontend with support for PyTorch☆25Updated last year