onnx / digestaiLinks
Digest AI is a powerful model analysis tool that extracts insights from your models.
☆29Updated last month
Alternatives and similar repositories for digestai
Users that are interested in digestai are comparing it to the libraries listed below
Sorting:
- Efficient in-memory representation for ONNX, in Python☆20Updated this week
- Home for OctoML PyTorch Profiler☆113Updated 2 years ago
- An experimental CPU backend for Triton (https//github.com/openai/triton)☆43Updated 4 months ago
- No-code CLI designed for accelerating ONNX workflows☆201Updated last month
- Notes and artifacts from the ONNX steering committee☆26Updated last week
- A fork of tvm/unity☆14Updated last year
- TORCH_LOGS parser for PT2☆46Updated this week
- TritonParse is a tool designed to help developers analyze and debug Triton kernels by visualizing the compilation process and source code…☆131Updated this week
- Model compression for ONNX☆96Updated 8 months ago
- Machine Learning Agility (MLAgility) benchmark and benchmarking tools☆39Updated 2 months ago
- MLIR-based partitioning system☆103Updated this week
- ☆35Updated this week
- A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.☆187Updated this week
- Explore training for quantized models☆20Updated this week
- An IR for efficiently simulating distributed ML computation.☆28Updated last year
- Development repository for the Triton language and compiler☆125Updated this week
- ☆48Updated this week
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆87Updated this week
- Ahead of Time (AOT) Triton Math Library☆70Updated this week
- ☆120Updated last year
- Scoreboard for ONNX Backend Compatibility☆29Updated this week
- OpenAI Triton backend for Intel® GPUs☆193Updated this week
- Intel® Extension for DeepSpeed* is an extension to DeepSpeed that brings feature support with SYCL kernels on Intel GPU(XPU) device. Note…☆61Updated 2 weeks ago
- AI Tensor Engine for ROCm☆232Updated this week
- ☆16Updated 2 weeks ago
- ☆69Updated 2 years ago
- Benchmarks to capture important workloads.☆31Updated 5 months ago
- ☆28Updated 2 weeks ago
- High-speed GEMV kernels, at most 2.7x speedup compared to pytorch baseline.☆112Updated last year
- 🚀 Collection of components for development, training, tuning, and inference of foundation models leveraging PyTorch native components.☆206Updated this week