cap-lab / jediLinks
Jetson embedded platform-target deep learning inference acceleration framework with TensorRT
☆29Updated last month
Alternatives and similar repositories for jedi
Users that are interested in jedi are comparing it to the libraries listed below
Sorting:
- Inference of quantization aware trained networks using TensorRT☆83Updated 2 years ago
- NVIDIA DLA-SW, the recipes and tools for running deep learning workloads on NVIDIA DLA cores for inference applications.☆219Updated last year
- A tutorial for getting started with the Deep Learning Accelerator (DLA) on NVIDIA Jetson☆355Updated 3 years ago
- A set of examples around MegEngine☆31Updated last year
- Offline Quantization Tools for Deploy.☆141Updated last year
- ☆163Updated 2 years ago
- Count number of parameters / MACs / FLOPS for ONNX models.☆94Updated last year
- ☆37Updated 2 years ago
- ☆38Updated last year
- Quick and Self-Contained TensorRT Custom Plugin Implementation and Integration☆72Updated 5 months ago
- Tencent Distribution of TVM☆15Updated 2 years ago
- ☆21Updated 4 years ago
- ☆60Updated 11 months ago
- Code for ACM MobiCom 2024 paper "FlexNN: Efficient and Adaptive DNN Inference on Memory-Constrained Edge Devices"☆56Updated 9 months ago
- ☆18Updated this week
- ☆98Updated 4 years ago
- Common libraries for PPL projects☆30Updated 8 months ago
- YOLOv5 on Orin DLA☆215Updated last year
- PyTorch Quantization Aware Training Example☆144Updated last year
- This is 8-bit quantization sample for yolov5. Both PTQ, QAT and Partial Quantization have been implemented, and present the results based…☆109Updated 3 years ago
- Fast CUDA Kernels for ResNet Inference.☆181Updated 6 years ago
- ☆19Updated last year
- Based of paper "Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference"☆64Updated 4 years ago
- ☆69Updated 3 years ago
- heterogeneity-aware-lowering-and-optimization☆256Updated last year
- ☆11Updated 10 months ago
- ☆243Updated 3 years ago
- [MLSys 2021] IOS: Inter-Operator Scheduler for CNN Acceleration☆200Updated 3 years ago
- Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. Th…☆424Updated this week
- VeriSilicon Tensor Interface Module☆239Updated last month