lvyufeng / easy_mindspore_bk
☆18Updated 2 years ago
Alternatives and similar repositories for easy_mindspore_bk:
Users that are interested in easy_mindspore_bk are comparing it to the libraries listed below
- ☆13Updated 2 years ago
- A Tight-fisted Optimizer☆47Updated last year
- [KDD'22] Learned Token Pruning for Transformers☆96Updated last year
- [EVA ICLR'23; LARA ICML'22] Efficient attention mechanisms via control variates, random features, and importance sampling☆80Updated last year
- [AAAI 2024] Fluctuation-based Adaptive Structured Pruning for Large Language Models☆42Updated last year
- [ICLR 2024] This is the official PyTorch implementation of "QLLM: Accurate and Efficient Low-Bitwidth Quantization for Large Language Mod…☆35Updated 10 months ago
- ☆35Updated 2 years ago
- ☆29Updated 2 months ago
- [ICLR 2022] "Unified Vision Transformer Compression" by Shixing Yu*, Tianlong Chen*, Jiayi Shen, Huan Yuan, Jianchao Tan, Sen Yang, Ji Li…☆52Updated last year
- [NeurIPS'21] "Chasing Sparsity in Vision Transformers: An End-to-End Exploration" by Tianlong Chen, Yu Cheng, Zhe Gan, Lu Yuan, Lei Zhang…☆90Updated last year
- Official implementation of the EMNLP23 paper: Outlier Suppression+: Accurate quantization of large language models by equivalent and opti…☆45Updated last year
- [TMLR] Official PyTorch implementation of paper "Efficient Quantization-aware Training with Adaptive Coreset Selection"☆30Updated 5 months ago
- The official implementation of the ICML 2023 paper OFQ-ViT☆30Updated last year
- Pytorch implementation of our paper accepted by ICML 2023 -- "Bi-directional Masks for Efficient N:M Sparse Training"☆11Updated last year
- 基于Gated Attention Unit的Transformer模型(尝鲜版)☆97Updated last year
- Official PyTorch implementation of IntactKV: Improving Large Language Model Quantization by Keeping Pivot Tokens Intact☆38Updated 8 months ago
- Must-read papers on improving efficiency for pre-trained language models.☆102Updated 2 years ago
- (CVPR 2022) Automated Progressive Learning for Efficient Training of Vision Transformers☆25Updated 2 years ago
- ☆11Updated last year
- ☆54Updated 5 months ago
- ☆59Updated 6 months ago
- [NeurIPS 2024] Search for Efficient LLMs☆12Updated 2 weeks ago
- Official Pytorch Implementation of Our Paper Accepted at ICLR 2024-- Dynamic Sparse No Training: Training-Free Fine-tuning for Sparse LLM…☆38Updated 9 months ago
- # Unified Normalization (ACM MM'22) By Qiming Yang, Kai Zhang, Chaoxiang Lan, Zhi Yang, Zheyang Li, Wenming Tan, Jun Xiao, and Shiliang P…☆34Updated last year
- [ICLR 2022] Official implementation of cosformer-attention in cosFormer: Rethinking Softmax in Attention☆185Updated 2 years ago
- https://hyperbox-doc.readthedocs.io/en/latest/☆25Updated last year
- Source code for IJCAI 2022 Long paper: Parameter-Efficient Sparsity for Large Language Models Fine-Tuning.☆13Updated 2 years ago
- BESA is a differentiable weight pruning technique for large language models.☆14Updated 10 months ago
- Official implement of Evo-ViT: Slow-Fast Token Evolution for Dynamic Vision Transformer☆72Updated 2 years ago
- Implementation of AAAI 2022 Paper: Go wider instead of deeper☆32Updated 2 years ago