bwconrad / soft-moe
PyTorch implementation of "From Sparse to Soft Mixtures of Experts"
☆50Updated last year
Alternatives and similar repositories for soft-moe:
Users that are interested in soft-moe are comparing it to the libraries listed below
- PyTorch implementation of Soft MoE by Google Brain in "From Sparse to Soft Mixtures of Experts" (https://arxiv.org/pdf/2308.00951.pdf)☆71Updated last year
- [NeurIPS 2023] Make Your Pre-trained Model Reversible: From Parameter to Memory Efficient Fine-Tuning☆30Updated last year
- [ICLR 2023] "Sparse MoE as the New Dropout: Scaling Dense and Self-Slimmable Transformers" by Tianlong Chen*, Zhenyu Zhang*, Ajay Jaiswal…☆48Updated last year
- ☆78Updated last year
- On the Effectiveness of Parameter-Efficient Fine-Tuning☆38Updated last year
- [EMNLP 2023 Main] Sparse Low-rank Adaptation of Pre-trained Language Models☆71Updated 11 months ago
- [EVA ICLR'23; LARA ICML'22] Efficient attention mechanisms via control variates, random features, and importance sampling☆80Updated last year
- Mixture of Attention Heads☆41Updated 2 years ago
- [CVPR 2022] "The Principle of Diversity: Training Stronger Vision Transformers Calls for Reducing All Levels of Redundancy" by Tianlong C…☆25Updated 2 years ago
- MambaFormer in-context learning experiments and implementation for https://arxiv.org/abs/2402.04248☆36Updated 7 months ago
- A repository for DenseSSMs☆86Updated 10 months ago
- This repository is the implementation of the paper Training Free Pretrained Model Merging (CVPR2024).☆27Updated 11 months ago
- EfficientVLM: Fast and Accurate Vision-Language Models via Knowledge Distillation and Modal-adaptive Pruning (ACL 2023)☆24Updated last year
- PyTorch implementation of LIMoE☆53Updated 10 months ago
- The this is the official implementation of "DAPE: Data-Adaptive Positional Encoding for Length Extrapolation"☆35Updated 4 months ago
- Repository containing code for blockwise SSL training☆28Updated 4 months ago
- ☆22Updated 4 months ago
- [NeurIPS 2024] For paper Parameter Competition Balancing for Model Merging☆33Updated 4 months ago
- Official code for "pi-Tuning: Transferring Multimodal Foundation Models with Optimal Multi-task Interpolation", ICML 2023.☆32Updated last year
- ☆33Updated 6 months ago
- ☆21Updated 2 years ago
- ☆41Updated 3 weeks ago
- Metrics for "Beyond neural scaling laws: beating power law scaling via data pruning " (NeurIPS 2022 Outstanding Paper Award)☆55Updated last year
- Structured Pruning Adapters in PyTorch☆16Updated last year
- Official implementation for NeurIPS'23 paper "Geodesic Multi-Modal Mixup for Robust Fine-Tuning"☆30Updated 4 months ago
- HGRN2: Gated Linear RNNs with State Expansion☆52Updated 5 months ago
- AutoPEFT: Automatic Configuration Search for Parameter-Efficient Fine-Tuning (Zhou et al.; TACL 2024)☆46Updated 10 months ago
- ☆99Updated 11 months ago
- ☆17Updated last month
- Code accompanying the paper "Massive Activations in Large Language Models"☆140Updated 11 months ago