nod-ai / SRT
Nod.ai π¦ version of π» . You probably want to start at https://github.com/nod-ai/shark for the product and the upstream IREE repository for mainline development. This repository houses branches and configuration that aren't ready for commit upstream.
β106Updated 4 months ago
Alternatives and similar repositories for SRT:
Users that are interested in SRT are comparing it to the libraries listed below
- benchmarking some transformer deploymentsβ26Updated 2 years ago
- MLIR-based partitioning systemβ82Updated this week
- β51Updated 8 months ago
- Fast sparse deep learning on CPUsβ53Updated 2 years ago
- Customized matrix multiplication kernelsβ54Updated 3 years ago
- Training neural networks in TensorFlow 2.0 with 5x less memoryβ131Updated 3 years ago
- A thin, highly portable toolkit for efficiently compiling dense loop-based computation.β148Updated 2 years ago
- Unified compiler/runtime for interfacing with PyTorch Dynamo.β100Updated 2 months ago
- Ahead of Time (AOT) Triton Math Libraryβ63Updated 2 weeks ago
- Home for OctoML PyTorch Profilerβ113Updated 2 years ago
- β50Updated last year
- 3X speedup over Appleβs TensorFlow plugin by using Apache TVM on M1β136Updated 3 years ago
- Benchmarks to capture important workloads.β31Updated 3 months ago
- Torch Distributed Experimentalβ115Updated 9 months ago
- β13Updated 3 years ago
- A library of GPU kernels for sparse matrix operations.β264Updated 4 years ago
- A stand-alone implementation of several NumPy dtype extensions used in machine learning.β261Updated this week
- torch::deploy (multipy for non-torch uses) is a system that lets you get around the GIL problem by running multiple Python interpreters iβ¦β180Updated 4 months ago
- extensible collectives library in tritonβ86Updated last month
- TileFusion is an experimental C++ macro kernel template library that elevates the abstraction level in CUDA C for tile processing.β84Updated last week
- A library for syntactically rewriting Python programs, pronounced (sinner).β69Updated 3 years ago
- A tracing JIT compiler for PyTorchβ13Updated 3 years ago
- A Data-Centric Compiler for Machine Learningβ82Updated last year
- β16Updated 7 months ago
- MatMul Performance Benchmarks for a Single CPU Core comparing both hand engineered and codegen kernels.β130Updated last year
- High-Performance SGEMM on CUDA devicesβ90Updated 3 months ago
- Codebase associated with the PyTorch compiler tutorialβ45Updated 5 years ago
- Memory Optimizations for Deep Learning (ICML 2023)β64Updated last year
- A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.β102Updated this week
- High-speed GEMV kernels, at most 2.7x speedup compared to pytorch baseline.β106Updated 9 months ago