jinqixiao / ComCATLinks
☆16Updated last year
Alternatives and similar repositories for ComCAT
Users that are interested in ComCAT are comparing it to the libraries listed below
Sorting:
- This repository is the implementation of the paper Training Free Pretrained Model Merging (CVPR2024).☆31Updated last year
- [NeurIPS 2022 Spotlight] This is the official PyTorch implementation of "EcoFormer: Energy-Saving Attention with Linear Complexity"☆74Updated 2 years ago
- Minimal multi-gpu implementation of EDM2: "Analyzing and Improving the Training Dynamics of Diffusion Models"☆34Updated last year
- VidKV: Plug-and-Play 1.x-Bit KV Cache Quantization for Video Large Language Models☆22Updated 5 months ago
- This is the official PyTorch implementation for "Mesa: A Memory-saving Training Framework for Transformers".☆120Updated 3 years ago
- [ICLR 2022] "As-ViT: Auto-scaling Vision Transformers without Training" by Wuyang Chen, Wei Huang, Xianzhi Du, Xiaodan Song, Zhangyang Wa…☆76Updated 3 years ago
- official code for Diff-Instruct algorithm for one-step diffusion distillation☆78Updated 7 months ago
- ☆52Updated last year
- ☆38Updated last year
- Diffusion Model as a Noise-Aware Latent Reward Model for Step-Level Preference Optimization☆38Updated 3 months ago
- [ICLR 2024 Spotlight] This is the official PyTorch implementation of "EfficientDM: Efficient Quantization-Aware Fine-Tuning of Low-Bit Di…☆65Updated last year
- Triton implement of bi-directional (non-causal) linear attention☆54Updated 7 months ago
- AliTok: Towards Sequence Modeling Alignment between Tokenizer and Autoregressive Model☆44Updated 2 months ago
- Implementation for <Orthogonal Over-Parameterized Training> in CVPR'21.☆21Updated 4 years ago
- TerDiT: Ternary Diffusion Models with Transformers☆71Updated last year
- ☆45Updated last year
- ☆28Updated last year
- [Interspeech 2024] LiteFocus is a tool designed to accelerate diffusion-based TTA model, now implemented with the base model AudioLDM2.☆33Updated 5 months ago
- Original code base for On Pretraining Data Diversity for Self-Supervised Learning☆14Updated 8 months ago
- [NeurIPS 2024] Exploring DCN-like Architectures for Fast Image Generation with Arbitrary Resolution☆34Updated 8 months ago
- ☆16Updated 2 years ago
- (CVPR 2022) Automated Progressive Learning for Efficient Training of Vision Transformers☆25Updated 6 months ago
- Are gradient information useful for pruning of LLMs?☆46Updated 2 weeks ago
- The official implementation of the paper "Reducing Fine-Tuning Memory Overhead by Approximate and Memory-Sharing Backpropagation"☆20Updated 8 months ago
- Official implementation of "Parameter-Efficient Orthogonal Finetuning via Butterfly Factorization"☆79Updated last year
- ☆56Updated last year
- The official PyTorch implementation of Fast Diffusion Model☆95Updated 2 years ago
- [Preprint] Why is the State of Neural Network Pruning so Confusing? On the Fairness, Comparison Setup, and Trainability in Network Prunin…☆40Updated 2 years ago
- [NeurIPS 2024] Learning-to-Cache: Accelerating Diffusion Transformer via Layer Caching☆112Updated last year
- [NeurIPS 2024] VeLoRA : Memory Efficient Training using Rank-1 Sub-Token Projections☆21Updated 10 months ago