ultralytics / thopLinks
Profile PyTorch models for FLOPs and parameters, helping to evaluate computational efficiency and memory usage.
☆65Updated 2 weeks ago
Alternatives and similar repositories for thop
Users that are interested in thop are comparing it to the libraries listed below
Sorting:
- Recent Advances on Efficient Vision Transformers☆54Updated 2 years ago
- [CVPR 2024] PTQ4SAM: Post-Training Quantization for Segment Anything☆82Updated last year
- When it comes to optimizers, it's always better to be safe than sorry☆378Updated last month
- A library for calculating the FLOPs in the forward() process based on torch.fx☆130Updated 7 months ago
- Easily benchmark PyTorch model FLOPs, latency, throughput, allocated gpu memory and energy consumption☆109Updated 2 years ago
- FlashFFTConv: Efficient Convolutions for Long Sequences with Tensor Cores☆331Updated 10 months ago
- [NeurIPS 2023] MCUFormer: Deploying Vision Transformers on Microcontrollers with Limited Memory☆74Updated 2 years ago
- [ECCV 2024] Isomorphic Pruning for Vision Models☆79Updated last year
- OTOv1-v3, NeurIPS, ICLR, TMLR, DNN Training, Compression, Structured Pruning, Erasing Operators, CNN, Diffusion, LLM☆309Updated last year
- Implementation of Post-training Quantization on Diffusion Models (CVPR 2023)☆139Updated 2 years ago
- Timm model explorer☆42Updated last year
- Awesome list of papers that extend Mamba to various applications.☆138Updated 5 months ago
- [ICML 2023] This project is the official implementation of our accepted ICML 2023 paper BiBench: Benchmarking and Analyzing Network Binar…☆55Updated last year
- [CVPR'23] SparseViT: Revisiting Activation Sparsity for Efficient High-Resolution Vision Transformer☆75Updated last year
- ☆186Updated last year
- Awesome Pruning. ✅ Curated Resources for Neural Network Pruning.☆170Updated last year
- The official implementation of the NeurIPS 2022 paper Q-ViT.☆100Updated 2 years ago
- An efficient pytorch implementation of selective scan in one file, works with both cpu and gpu, with corresponding mathematical derivatio…☆97Updated last month
- [NeurIPS 2025 Spotlight] TPA: Tensor ProducT ATTenTion Transformer (T6) (https://arxiv.org/abs/2501.06425)☆425Updated 3 weeks ago
- ☆165Updated 2 years ago
- ☆61Updated last year
- Get down and dirty with FlashAttention2.0 in pytorch, plug in and play no complex CUDA kernels☆109Updated 2 years ago
- This library empowers users to seamlessly port pretrained models and checkpoints on the HuggingFace (HF) hub (developed using HF transfor…☆83Updated this week
- Code for the NeurIPS 2022 paper "Optimal Brain Compression: A Framework for Accurate Post-Training Quantization and Pruning".☆129Updated 2 years ago
- Official implementation of the EMNLP23 paper: Outlier Suppression+: Accurate quantization of large language models by equivalent and opti…☆47Updated 2 years ago
- A simple program to calculate and visualize the FLOPs and Parameters of Pytorch models, with handy CLI and easy-to-use Python API.☆131Updated 11 months ago
- OTOv1-v3, NeurIPS, ICLR, TMLR, DNN Training, Compression, Structured Pruning, Erasing Operators, CNN, Diffusion, LLM☆48Updated last year
- List of papers related to Vision Transformers quantization and hardware acceleration in recent AI conferences and journals.☆96Updated last year
- Speedup the attention computation of Swin Transformer☆25Updated 5 months ago
- [TMLR] Official PyTorch implementation of paper "Quantization Variation: A New Perspective on Training Transformers with Low-Bit Precisio…☆46Updated last year