CerebrasResearch / RevBiFPNLinks
RevBiFPN: The Fully Reversible Bidirectional Feature Pyramid Network
☆14Updated 2 years ago
Alternatives and similar repositories for RevBiFPN
Users that are interested in RevBiFPN are comparing it to the libraries listed below
Sorting:
- Seamless analysis of your PyTorch models (RAM usage, FLOPs, MACs, receptive field, etc.)☆218Updated 4 months ago
- Using ideas from product quantization for state-of-the-art neural network compression.☆145Updated 3 years ago
- A research library for pytorch-based neural network pruning, compression, and more.☆162Updated 2 years ago
- Transformers w/o Attention, based fully on MLPs☆93Updated last year
- A implement of run-length encoding for Pytorch tensor using CUDA☆14Updated 4 years ago
- [NeurIPS 2022 Spotlight] This is the official PyTorch implementation of "EcoFormer: Energy-Saving Attention with Linear Complexity"☆72Updated 2 years ago
- ☆43Updated last year
- Implementation of fused cosine similarity attention in the same style as Flash Attention☆214Updated 2 years ago
- Estimate/count FLOPS for a given neural network using pytorch☆305Updated 3 years ago
- A better PyTorch implementation of image local attention which reduces the GPU memory by an order of magnitude.☆141Updated 3 years ago
- MONeT framework for reducing memory consumption of DNN training☆173Updated 4 years ago
- DeltaCNN End-to-End CNN Inference of Sparse Frame Differences in Videos☆59Updated 2 years ago
- Code for ICML 2022 paper "SPDY: Accurate Pruning with Speedup Guarantees"☆20Updated 2 years ago
- AlphaNet Improved Training of Supernet with Alpha-Divergence☆99Updated 3 years ago
- Neural Architecture Search for Neural Network Libraries☆60Updated last year
- Nested Hierarchical Transformer https://arxiv.org/pdf/2105.12723.pdf☆198Updated 11 months ago
- Simple CIFAR-10 classification with ConvMixer☆45Updated 3 years ago
- Simplification of pruned models for accelerated inference | SoftwareX https://doi.org/10.1016/j.softx.2021.100907☆36Updated 5 months ago
- Is the attention layer even necessary? (https://arxiv.org/abs/2105.02723)☆486Updated 4 years ago
- A crash course on PyTorch hooks☆40Updated 5 years ago
- [ECCV 2022] SuperTickets: Drawing Task-Agnostic Lottery Tickets from Supernets via Jointly Architecture Searching and Parameter Pruning☆21Updated 3 years ago
- Official PyTorch Implementation of HELP: Hardware-adaptive Efficient Latency Prediction for NAS via Meta-Learning (NeurIPS 2021 Spotlight…☆63Updated 11 months ago
- PyTorch implementation of LARS (Layer-wise Adaptive Rate Scaling)☆20Updated 6 years ago
- ☆69Updated 5 years ago
- TF/Keras code for DiffStride, a pooling layer with learnable strides.☆124Updated 3 years ago
- [ICML 2022] "DepthShrinker: A New Compression Paradigm Towards Boosting Real-Hardware Efficiency of Compact Neural Networks", by Yonggan …