google-ai-edge / ai-edge-quantizerLinks
AI Edge Quantizer: flexible post training quantization for LiteRT models.
☆72Updated this week
Alternatives and similar repositories for ai-edge-quantizer
Users that are interested in ai-edge-quantizer are comparing it to the libraries listed below
Sorting:
- Model compression for ONNX☆97Updated 11 months ago
- This library empowers users to seamlessly port pretrained models and checkpoints on the HuggingFace (HF) hub (developed using HF transfor…☆82Updated this week
- 🤗 Optimum ExecuTorch☆74Updated this week
- Inference Vision Transformer (ViT) in plain C/C++ with ggml☆295Updated last year
- ONNX and TensorRT implementation of Whisper☆64Updated 2 years ago
- Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. Th…☆419Updated last week
- A Toolkit to Help Optimize Onnx Model☆228Updated this week
- Convert tflite to JSON and make it editable in the IDE. It also converts the edited JSON back to tflite binary.☆27Updated 2 years ago
- New operators for the ReferenceEvaluator, new kernels for onnxruntime, CPU, CUDA☆35Updated last week
- Visualize ONNX models with model-explorer☆59Updated 2 weeks ago
- Mobile App Open☆63Updated this week
- A very simple tool for situations where optimization with onnx-simplifier would exceed the Protocol Buffers upper file size limit of 2GB,…☆17Updated 3 weeks ago
- Use safetensors with ONNX 🤗☆73Updated 3 weeks ago
- ONNX implementation of Whisper. PyTorch free.☆101Updated 11 months ago
- QONNX: Arbitrary-Precision Quantized Neural Networks in ONNX☆161Updated this week
- Inference RWKV v5, v6 and v7 with Qualcomm AI Engine Direct SDK