amazon-science / adaptive-feature-transferLinks
Official implementation of Adaptive Feature Transfer (AFT)
☆23Updated last year
Alternatives and similar repositories for adaptive-feature-transfer
Users that are interested in adaptive-feature-transfer are comparing it to the libraries listed below
Sorting:
- HGRN2: Gated Linear RNNs with State Expansion☆54Updated last year
- Official code for the paper "Attention as a Hypernetwork"☆42Updated last year
- Code and benchmark for the paper: "A Practitioner's Guide to Continual Multimodal Pretraining" [NeurIPS'24]☆58Updated 9 months ago
- Official implementation of RMoE (Layerwise Recurrent Router for Mixture-of-Experts)☆23Updated last year
- User-friendly implementation of the Mixture-of-Sparse-Attention (MoSA). MoSA selects distinct tokens for each head with expert choice rou…☆26Updated 4 months ago
- This is a PyTorch implementation of the paperViP A Differentially Private Foundation Model for Computer Vision☆36Updated 2 years ago
- Official PyTorch Implementation for Vision-Language Models Create Cross-Modal Task Representations, ICML 2025☆31Updated 4 months ago
- Official Code for ICLR 2024 Paper: Non-negative Contrastive Learning☆45Updated last year
- [ICLR 2025 & COLM 2025] Official PyTorch implementation of the Forgetting Transformer and Adaptive Computation Pruning☆132Updated last week
- [ICML 2025] Code for "R2-T2: Re-Routing in Test-Time for Multimodal Mixture-of-Experts"☆16Updated 6 months ago
- Why Do We Need Weight Decay in Modern Deep Learning? [NeurIPS 2024]☆67Updated last year
- State Space Models☆70Updated last year
- Autoregressive Image Generation☆32Updated 3 months ago
- An official PyTorch implementation for CLIPPR☆29Updated 2 years ago
- ☆32Updated 7 months ago
- [ICML 2024] When Linear Attention Meets Autoregressive Decoding: Towards More Effective and Efficient Linearized Large Language Models☆33Updated last year
- Explorations into adversarial losses on top of autoregressive loss for language modeling☆37Updated 7 months ago
- Official repository of "LiNeS: Post-training Layer Scaling Prevents Forgetting and Enhances Model Merging"☆30Updated 10 months ago
- Implementation of Agent Attention in Pytorch☆91Updated last year
- This is a simple torch implementation of the high performance Multi-Query Attention☆16Updated 2 years ago
- Latest Weight Averaging (NeurIPS HITY 2022)☆31Updated 2 years ago
- More dimensions = More fun☆25Updated last year
- ☆35Updated 6 months ago
- PyTorch implementation of Soft MoE by Google Brain in "From Sparse to Soft Mixtures of Experts" (https://arxiv.org/pdf/2308.00951.pdf)☆78Updated last year
- Unofficial Implementation of Selective Attention Transformer☆17Updated 10 months ago
- ☆58Updated 11 months ago
- ☆47Updated last year
- The open-source materials for paper "Sparsing Law: Towards Large Language Models with Greater Activation Sparsity".☆25Updated 10 months ago
- Official repository of paper "RNNs Are Not Transformers (Yet): The Key Bottleneck on In-context Retrieval"☆27Updated last year
- Here we will test various linear attention designs.☆62Updated last year