lvyufeng / easy_mindspore_bk
☆18Updated 2 years ago
Alternatives and similar repositories for easy_mindspore_bk
Users that are interested in easy_mindspore_bk are comparing it to the libraries listed below
Sorting:
- [KDD'22] Learned Token Pruning for Transformers☆97Updated 2 years ago
- ☆13Updated 2 years ago
- Official PyTorch implementation of "IntactKV: Improving Large Language Model Quantization by Keeping Pivot Tokens Intact"☆44Updated 11 months ago
- A Tight-fisted Optimizer☆47Updated 2 years ago
- The official implementation of the ICML 2023 paper OFQ-ViT☆30Updated last year
- qwen-nsa☆61Updated last month
- Official implementation of the EMNLP23 paper: Outlier Suppression+: Accurate quantization of large language models by equivalent and opti…☆46Updated last year
- [AAAI 2024] Fluctuation-based Adaptive Structured Pruning for Large Language Models☆49Updated last year
- Official code of *Virgo: A Preliminary Exploration on Reproducing o1-like MLLM*☆20Updated 2 months ago
- mindspore implementation of transformers☆66Updated 2 years ago
- The official PyTorch implementation of the NeurIPS2022 (spotlight) paper, Outlier Suppression: Pushing the Limit of Low-bit Transformer L…☆47Updated 2 years ago
- [ICLR 2024] This is the official PyTorch implementation of "QLLM: Accurate and Efficient Low-Bitwidth Quantization for Large Language Mod…☆36Updated last year
- [EVA ICLR'23; LARA ICML'22] Efficient attention mechanisms via control variates, random features, and importance sampling☆85Updated 2 years ago
- ☆19Updated 4 months ago
- ☆11Updated last year
- (CVPR 2022) Automated Progressive Learning for Efficient Training of Vision Transformers☆25Updated 2 months ago
- Source code for IJCAI 2022 Long paper: Parameter-Efficient Sparsity for Large Language Models Fine-Tuning.☆14Updated 2 years ago
- ☆20Updated 6 months ago
- Due to the huge vocaburary size (151,936) of Qwen models, the Embedding and LM Head weights are excessively heavy. Therefore, this projec…☆20Updated 9 months ago
- Summary of system papers/frameworks/codes/tools on training or serving large model☆56Updated last year
- ☆55Updated 8 months ago
- [ICLR 2022] "Unified Vision Transformer Compression" by Shixing Yu*, Tianlong Chen*, Jiayi Shen, Huan Yuan, Jianchao Tan, Sen Yang, Ji Li…☆52Updated last year
- [ICLR 2025] The official pytorch implement of "Dynamic Low-Rank Sparse Adaptation for Large Language Models".☆18Updated 2 months ago
- [NeurIPS'21] "Chasing Sparsity in Vision Transformers: An End-to-End Exploration" by Tianlong Chen, Yu Cheng, Zhe Gan, Lu Yuan, Lei Zhang…☆89Updated last year
- Less is More: Task-aware Layer-wise Distillation for Language Model Compression (ICML2023)☆34Updated last year
- ☆14Updated last year
- official code for GliDe with a CaPE☆19Updated 9 months ago
- BESA is a differentiable weight pruning technique for large language models.☆16Updated last year
- Implementation of AAAI 2022 Paper: Go wider instead of deeper☆32Updated 2 years ago
- ☆49Updated last year