OAID / AutoKernelLinks
AutoKernel 是一个简单易用,低门槛的自动算子优化工具,提高深度学习算法部署效率。
☆742Updated 3 years ago
Alternatives and similar repositories for AutoKernel
Users that are interested in AutoKernel are comparing it to the libraries listed below
Sorting:
- ☆253Updated 2 years ago
- 🔥 (yolov3 yolov4 yolov5 unet ...)A mini pytorch inference framework which inspired from darknet.☆746Updated 2 years ago
- benchmark for embededded-ai deep learning inference engines, such as NCNN / TNN / MNN / TensorFlow Lite etc.☆204Updated 4 years ago
- Tengine Convert Tool supports converting multi framworks' models into tmfile that suitable for Tengine-Lite AI framework.☆92Updated 4 years ago
- ppl.cv is a high-performance image processing library of openPPL supporting various platforms.☆511Updated 11 months ago
- MegCC是一个运行时超轻量,高效,移植简单的深度学习模型编译器☆486Updated last year
- ☆98Updated 4 years ago
- SuperSonic, a new open-source framework to allow compiler developers to integrate RL into compilers easily, regardless of their RL expert…☆122Updated 2 years ago
- Bolt is a deep learning library with high performance and heterogeneous flexibility.☆953Updated 6 months ago
- A library for high performance deep learning inference on NVIDIA GPUs.☆557Updated 3 years ago
- Compiler Infrastructure for Neural Networks☆147Updated 2 years ago
- mperf是一个面向移动/嵌入式平台的算子性能调优工具箱☆190Updated 2 years ago
- EasyQuant(EQ) is an efficient and simple post-training quantization method via effectively optimizing the scales of weights and activatio…☆405Updated 2 years ago
- VeriSilicon Tensor Interface Module☆238Updated 2 weeks ago
- row-major matmul optimization☆682Updated 2 months ago
- A primitive library for neural network☆1,364Updated 11 months ago
- High performance Cross-platform Inference-engine, you could run Anakin on x86-cpu,arm, nv-gpu, amd-gpu,bitmain and cambricon devices.☆534Updated 3 years ago
- arm-neon☆92Updated last year
- caffe model convert to onnx model☆176Updated 2 years ago
- 动手学习TVM核心原理教程☆63Updated 4 years ago
- symmetric int8 gemm☆67Updated 5 years ago
- heterogeneity-aware-lowering-and-optimization☆256Updated last year
- An acceleration library that supports arbitrary bit-width combinatorial quantization operations☆236Updated last year
- FlagPerf is an open-source software platform for benchmarking AI chips.☆352Updated 2 weeks ago
- Higher performance OpenAI LLM service than vLLM serve: A pure C++ high-performance OpenAI LLM service implemented with GPRS+TensorRT-LLM+…☆156Updated 5 months ago
- CNStream is a streaming framework for building Cambricon machine learning pipelines http://forum.cambricon.com https://gitee.com/Solu…☆53Updated 7 months ago
- TensorRT Plugin Autogen Tool☆368Updated 2 years ago
- Deep Learning Deployment Framework: Supports tf/torch/trt/trtllm/vllm and other NN frameworks. Support dynamic batching, and streaming mo…☆167Updated 5 months ago
- Tengine gemm tutorial, step by step☆13Updated 4 years ago
- Generate a quantization parameter file for ncnn framework int8 inference☆517Updated 5 years ago