[CVPR 2023] Castling-ViT: Compressing Self-Attention via Switching Towards Linear-Angular Attention During Vision Transformer Inference
☆30Mar 14, 2024Updated last year
Alternatives and similar repositories for Castling-ViT
Users that are interested in Castling-ViT are comparing it to the libraries listed below
Sorting:
- [ICML 2024] When Linear Attention Meets Autoregressive Decoding: Towards More Effective and Efficient Linearized Large Language Models☆35Jun 12, 2024Updated last year
- ☆10Dec 13, 2022Updated 3 years ago
- TiC: Exploring Vision Transformer in Convolution☆11Oct 24, 2023Updated 2 years ago
- Official Implementation of the CVPR'23 paper 'Regularization of polynomial networks for image recognition'.☆10Jun 8, 2023Updated 2 years ago
- [NeurIPS 2022 Spotlight] This is the official PyTorch implementation of "EcoFormer: Energy-Saving Attention with Linear Complexity"☆74Nov 15, 2022Updated 3 years ago
- Code Repository for the NeurIPS 2022 paper: "Hyper-Representations as Generative Models: Sampling Unseen Neural Network Weights".☆18Jul 10, 2024Updated last year
- Implementation for paper Automata Extraction from Transformers.☆12Jun 8, 2024Updated last year
- Arrow Matrix Decomposition - Communication-Efficient Distributed Sparse Matrix Multiplication☆15Mar 25, 2024Updated last year
- Code for the paper "Stack Attention: Improving the Ability of Transformers to Model Hierarchical Patterns"