apple / ml-batchquantLinks
☆23Updated 3 years ago
Alternatives and similar repositories for ml-batchquant
Users that are interested in ml-batchquant are comparing it to the libraries listed below
Sorting:
- This repository contains the official implementation for the ECCV'22 paper, "SPIN: An Empirical Evaluation on Sharing Parameters of Isotr…☆20Updated 2 years ago
- ☆14Updated last month
- Export utility for unconstrained channel pruned models☆72Updated 2 years ago
- Self-Conditioning Pre-Trained Language Models, ICML 2022☆34Updated 3 years ago
- ☆42Updated 3 years ago
- ☆19Updated 4 years ago
- Repository accompanying the Interspeech 2022 publication titled "Space-Efficient Representation of Entity-centric Query Language Models" …☆13Updated 3 years ago
- A light-weight implementation of ICCV2023 paper "Reinforce Data, Multiply Impact: Improved Model Accuracy and Robustness with Dataset Rei…☆83Updated 2 years ago
- Research publication code for "Forward Compatible Training for Large-Scale Embedding Retrieval Systems", CVPR 2022, and "FastFill: Effici…☆56Updated 2 years ago
- ☆91Updated last year
- Research publication code for "Least Squares Binary Quantization of Neural Networks"☆83Updated 2 years ago
- ☆23Updated 3 years ago
- ☆48Updated 2 years ago
- Efficient in-memory representation for ONNX, in Python☆34Updated this week
- ☆13Updated last year
- torchvision-based transforms that provide access to parameterization☆15Updated last week
- Neural Architecture Search for Neural Network Libraries☆60Updated last year
- ☆15Updated 2 years ago
- Prototype routines for GPU quantization written using PyTorch.☆21Updated 4 months ago
- A block oriented training approach for inference time optimization.☆33Updated last year
- ptq4vm official repository☆24Updated 8 months ago
- ACL 2023☆39Updated 2 years ago
- AI Edge Quantizer: flexible post training quantization for LiteRT models.☆82Updated this week
- Flexible simulator for mixed precision and format simulation of LLMs and vision transformers.☆51Updated 2 years ago
- See the device (CPU/GPU/ANE) and estimated cost for every layer in your CoreML model.☆24Updated last month
- ShiftAddLLM: Accelerating Pretrained LLMs via Post-Training Multiplication-Less Reparameterization☆111Updated last year
- Open Source Projects from Pallas Lab☆21Updated 4 years ago
- MLPerf Mobile benchmarks☆14Updated 6 months ago
- This is a collection of our research on efficient AI, covering hardware-aware NAS and model compression.☆84Updated last year
- ☆26Updated 2 years ago