cap-lab / jedi
Jetson embedded platform-target deep learning inference acceleration framework with TensorRT
☆24Updated last week
Related projects ⓘ
Alternatives and complementary repositories for jedi
- Inference of quantization aware trained networks using TensorRT☆79Updated last year
- Offline Quantization Tools for Deploy.☆116Updated 10 months ago
- A tutorial for getting started with the Deep Learning Accelerator (DLA) on NVIDIA Jetson☆287Updated 2 years ago
- Implementation of YOLOv9 QAT optimized for deployment on TensorRT platforms.☆81Updated last week
- NVIDIA DLA-SW, the recipes and tools for running deep learning workloads on NVIDIA DLA cores for inference applications.☆179Updated 5 months ago
- ☆56Updated 2 years ago
- Benchmark inference speed of CNNs with various quantization methods in Pytorch+TensorRT with Jetson Nano/Xavier☆54Updated last year
- Count number of parameters / MACs / FLOPS for ONNX models.☆88Updated 2 weeks ago
- This repository provides YOLOV5 GPU optimization sample☆100Updated last year
- A set of examples around MegEngine☆30Updated 11 months ago
- [MLSys 2021] IOS: Inter-Operator Scheduler for CNN Acceleration☆194Updated 2 years ago
- YOLOv5 on Orin DLA☆184Updated 8 months ago
- ☆67Updated last year
- ☆121Updated last year
- ☆17Updated 4 years ago
- tophub autotvm log collections☆70Updated last year
- Benchmark scripts for TVM☆73Updated 2 years ago
- ☆32Updated last month
- This is a list of awesome edgeAI inference related papers.☆88Updated 10 months ago
- A code generator from ONNX to PyTorch code☆132Updated last year
- A simple tool that can generate TensorRT plugin code quickly.☆221Updated last year
- study of cutlass☆19Updated this week
- ☆34Updated 2 years ago
- PyTorch Quantization Aware Training Example☆123Updated 5 months ago
- QONNX: Arbitrary-Precision Quantized Neural Networks in ONNX☆126Updated 2 weeks ago
- Collection of blogs on AI development☆14Updated 3 months ago
- This repository contains the results and code for the MLPerf™ Inference v1.0 benchmark.☆30Updated last year
- Quick and Self-Contained TensorRT Custom Plugin Implementation and Integration☆37Updated 4 months ago
- NART = NART is not A RunTime, a deep learning inference framework.☆38Updated last year
- ☆44Updated 3 years ago