huawei-noah / bolt
Bolt is a deep learning library with high performance and heterogeneous flexibility.
☆933Updated 5 months ago
Alternatives and similar repositories for bolt:
Users that are interested in bolt are comparing it to the libraries listed below
- ☆1,008Updated 10 months ago
- A library for high performance deep learning inference on NVIDIA GPUs.☆550Updated 2 years ago
- High performance Cross-platform Inference-engine, you could run Anakin on x86-cpu,arm, nv-gpu, amd-gpu,bitmain and cambricon devices.☆531Updated 2 years ago
- ppl.cv is a high-performance image processing library of openPPL supporting various platforms.☆495Updated 2 months ago
- A primitive library for neural network☆1,308Updated last month
- BladeDISC is an end-to-end DynamIc Shape Compiler project for machine learning workloads.☆834Updated 2 weeks ago
- Model Quantization Benchmark☆779Updated this week
- A flexible and efficient deep neural network (DNN) compiler that generates high-performance executable from a DNN model description.☆972Updated 3 months ago
- Quantized Neural Network PACKage - mobile-optimized implementation of quantized neural network operators☆1,533Updated 5 years ago
- EasyQuant(EQ) is an efficient and simple post-training quantization method via effectively optimizing the scales of weights and activatio…☆393Updated 2 years ago
- MegCC是一个运行时超轻量,高效,移植简单的深度学习模型编译器☆476Updated 2 months ago
- benchmark for embededded-ai deep learning inference engines, such as NCNN / TNN / MNN / TensorFlow Lite etc.☆204Updated 3 years ago
- 服务侧深度学习部署案例☆451Updated 4 years ago
- Benchmarking Neural Network Inference on Mobile Devices☆365Updated last year
- PPL Quantization Tool (PPQ) is a powerful offline neural network quantization tool.☆1,610Updated 9 months ago
- Dive into Deep Learning Compiler☆646Updated 2 years ago
- Deploy your model with TensorRT quickly.☆764Updated last year
- A parser, editor and profiler tool for ONNX models.☆411Updated last week
- row-major matmul optimization☆599Updated last year
- TVM integration into PyTorch☆453Updated 5 years ago
- MNN applications by MNN, JNI exec, RK3399. Support tflite\tensorflow\caffe\onnx models.☆503Updated 5 years ago
- Generate a quantization parameter file for ncnn framework int8 inference☆521Updated 4 years ago
- Simplify your onnx model☆3,942Updated 4 months ago
- The Tensor Algebra SuperOptimizer for Deep Learning☆696Updated last year
- Everything in Torch Fx☆341Updated 7 months ago
- ONNX-TensorRT: TensorRT backend for ONNX☆2,996Updated last month
- ☆663Updated 3 years ago
- TensorRT Plugin Autogen Tool☆369Updated last year
- TinyNeuralNetwork is an efficient and easy-to-use deep learning model compression framework.☆775Updated 3 weeks ago
- a fast and user-friendly runtime for transformer inference (Bert, Albert, GPT2, Decoders, etc) on CPU and GPU.☆1,503Updated last year