cap-lab / jedi
Jetson embedded platform-target deep learning inference acceleration framework with TensorRT
☆26Updated this week
Alternatives and similar repositories for jedi:
Users that are interested in jedi are comparing it to the libraries listed below
- Inference of quantization aware trained networks using TensorRT☆80Updated 2 years ago
- A tutorial for getting started with the Deep Learning Accelerator (DLA) on NVIDIA Jetson☆316Updated 2 years ago
- This is 8-bit quantization sample for yolov5. Both PTQ, QAT and Partial Quantization have been implemented, and present the results based…☆98Updated 2 years ago
- [MLSys 2021] IOS: Inter-Operator Scheduler for CNN Acceleration☆197Updated 2 years ago
- YOLOv5 on Orin DLA☆189Updated last year
- Quick and Self-Contained TensorRT Custom Plugin Implementation and Integration☆51Updated 8 months ago
- NVIDIA DLA-SW, the recipes and tools for running deep learning workloads on NVIDIA DLA cores for inference applications.☆189Updated 8 months ago
- PyTorch Static Quantization Example☆38Updated 3 years ago
- Count number of parameters / MACs / FLOPS for ONNX models.☆90Updated 3 months ago
- ☆17Updated 4 years ago
- Common libraries for PPL projects☆29Updated 4 months ago
- This repository provides YOLOV5 GPU optimization sample☆102Updated 2 years ago
- ☆69Updated last year
- Offline Quantization Tools for Deploy.☆123Updated last year
- ☆66Updated 2 years ago
- Based of paper "Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference"☆62Updated 4 years ago
- llama INT4 cuda inference with AWQ☆50Updated last month
- Low Precision(quantized) Yolov5☆33Updated last year
- Benchmark inference speed of CNNs with various quantization methods in Pytorch+TensorRT with Jetson Nano/Xavier☆55Updated last year
- This is a list of awesome edgeAI inference related papers.☆92Updated last year
- PyTorch emulation library for Microscaling (MX)-compatible data formats☆199Updated 4 months ago
- ☆136Updated last year
- ☆13Updated 3 years ago
- PyTorch Quantization Aware Training Example☆128Updated 9 months ago
- Several optimization methods of half-precision general matrix vector multiplication (HGEMV) using CUDA core.☆55Updated 5 months ago
- Implementation of YOLOv9 QAT optimized for deployment on TensorRT platforms.☆100Updated 3 months ago
- Code for ACM MobiCom 2024 paper "FlexNN: Efficient and Adaptive DNN Inference on Memory-Constrained Edge Devices"☆49Updated 3 weeks ago
- ☆35Updated 4 months ago
- ☆34Updated 2 years ago
- A standalone GEMM kernel for fp16 activation and quantized weight, extracted from FasterTransformer☆88Updated 11 months ago