cap-lab / jediLinks
Jetson embedded platform-target deep learning inference acceleration framework with TensorRT
☆29Updated 2 months ago
Alternatives and similar repositories for jedi
Users that are interested in jedi are comparing it to the libraries listed below
Sorting:
- Inference of quantization aware trained networks using TensorRT☆83Updated 2 years ago
- NVIDIA DLA-SW, the recipes and tools for running deep learning workloads on NVIDIA DLA cores for inference applications.☆224Updated last year
- Count number of parameters / MACs / FLOPS for ONNX models.☆95Updated last year
- Offline Quantization Tools for Deploy.☆141Updated last year
- A set of examples around MegEngine☆31Updated 2 years ago
- Code for ACM MobiCom 2024 paper "FlexNN: Efficient and Adaptive DNN Inference on Memory-Constrained Edge Devices"☆58Updated 11 months ago
- PyTorch Quantization Aware Training Example☆146Updated last year
- ☆168Updated 2 years ago
- MegEngine到其他框架的转换器☆69Updated 2 years ago
- A parser, editor and profiler tool for ONNX models.☆468Updated last month
- Tencent Distribution of TVM☆15Updated 2 years ago
- A tutorial for getting started with the Deep Learning Accelerator (DLA) on NVIDIA Jetson☆361Updated 3 years ago
- ☆17Updated 5 years ago
- ☆98Updated 4 years ago
- Quick and Self-Contained TensorRT Custom Plugin Implementation and Integration☆74Updated 7 months ago
- ☆37Updated 3 years ago
- FakeQuantize with Learned Step Size(LSQ+) as Observer in PyTorch☆37Updated 4 years ago
- Benchmark scripts for TVM☆74Updated 3 years ago
- A DNN inference latency prediction toolkit for accurately modeling and predicting the latency on diverse edge devices.☆360Updated last year
- A Toolkit to Help Optimize Large Onnx Model☆162Updated 2 months ago
- A simple tool that can generate TensorRT plugin code quickly.☆238Updated 2 years ago
- This is a list of awesome edgeAI inference related papers.☆97Updated 2 years ago
- ☆44Updated 4 years ago
- ☆68Updated 2 years ago
- Jetson Benchmark☆399Updated last year
- Fast CUDA Kernels for ResNet Inference.☆182Updated 6 years ago
- Tencent NCNN with added CUDA support☆71Updated 4 years ago
- Qualcomm Hexagon NN Offload Framework☆45Updated 5 years ago
- symmetric int8 gemm☆67Updated 5 years ago
- YOLOv5 on Orin DLA☆216Updated last year