SqueezeBits / owlite-examples
OwLite Examples repository offers illustrative example codes to help users seamlessly compress PyTorch deep learning models and transform them into TensorRT engines.
☆10Updated 5 months ago
Alternatives and similar repositories for owlite-examples:
Users that are interested in owlite-examples are comparing it to the libraries listed below
- OwLite is a low-code AI model compression toolkit for AI models.☆42Updated 3 weeks ago
- Code for the AAAI 2024 Oral paper "OWQ: Outlier-Aware Weight Quantization for Efficient Fine-Tuning and Inference of Large Language Model…☆58Updated last year
- Study Group of Deep Learning Compiler☆156Updated 2 years ago
- ☆56Updated 2 years ago
- ☆50Updated 3 months ago
- ☆52Updated 11 months ago
- QUICK: Quantization-aware Interleaving and Conflict-free Kernel for efficient LLM inference☆116Updated last year
- PyTorch CoreSIG☆56Updated 2 months ago
- [ICML 2024 Oral] Any-Precision LLM: Low-Cost Deployment of Multiple, Different-Sized LLMs☆97Updated 2 months ago
- LLMServingSim: A HW/SW Co-Simulation Infrastructure for LLM Inference Serving at Scale☆93Updated 2 weeks ago
- ☆83Updated 11 months ago
- ☆102Updated last year
- Official PyTorch Implementation of HELP: Hardware-adaptive Efficient Latency Prediction for NAS via Meta-Learning (NeurIPS 2021 Spotlight…☆62Updated 7 months ago
- BSQ: Exploring Bit-Level Sparsity for Mixed-Precision Neural Network Quantization (ICLR 2021)☆40Updated 4 years ago
- ☆25Updated 2 years ago
- ☆141Updated 2 years ago
- PyTorch emulation library for Microscaling (MX)-compatible data formats☆207Updated 5 months ago
- Study parallel programming - CUDA, OpenMP, MPI, Pthread☆56Updated 2 years ago
- [ICML'21 Oral] I-BERT: Integer-only BERT Quantization☆239Updated 2 years ago
- ☆202Updated 3 years ago
- A performance library for machine learning applications.☆183Updated last year
- This repository contains integer operators on GPUs for PyTorch.☆193Updated last year
- ☆47Updated 3 years ago
- ☆91Updated last year
- Tender: Accelerating Large Language Models via Tensor Decompostion and Runtime Requantization (ISCA'24)☆13Updated 8 months ago
- Official implementation of EMNLP'23 paper "Revisiting Block-based Quantisation: What is Important for Sub-8-bit LLM Inference?"☆19Updated last year
- List of papers related to Vision Transformers quantization and hardware acceleration in recent AI conferences and journals.☆76Updated 9 months ago
- Experimental deep learning framework written in Rust☆14Updated 2 years ago
- PyTorch extension for emulating FP8 data formats on standard FP32 Xeon/GPU hardware.☆107Updated 3 months ago