pixas / NoRMLinks
ICLR 2025
☆27Updated 2 months ago
Alternatives and similar repositories for NoRM
Users that are interested in NoRM are comparing it to the libraries listed below
Sorting:
- The this is the official implementation of "DAPE: Data-Adaptive Positional Encoding for Length Extrapolation"☆38Updated 9 months ago
- This repo contains the source code for VB-LoRA: Extreme Parameter Efficient Fine-Tuning with Vector Banks (NeurIPS 2024).☆39Updated 9 months ago
- Official code for our paper, "LoRA-Pro: Are Low-Rank Adapters Properly Optimized? "☆126Updated 3 months ago
- ☆21Updated 2 months ago
- ☆108Updated last year
- EMPO, A Fully Unsupervised RLVR Method☆51Updated last week
- [ICML 2024] CrossGET: Cross-Guided Ensemble of Tokens for Accelerating Vision-Language Transformers.☆33Updated 6 months ago
- CLIP-MoE: Mixture of Experts for CLIP☆42Updated 9 months ago
- Official code for "pi-Tuning: Transferring Multimodal Foundation Models with Optimal Multi-task Interpolation", ICML 2023.☆33Updated last year
- Representation Surgery for Multi-Task Model Merging. ICML, 2024.☆45Updated 9 months ago
- ☆42Updated 8 months ago
- [AAAI 2025] HiRED strategically drops visual tokens in the image encoding stage to improve inference efficiency for High-Resolution Visio…☆39Updated 3 months ago
- Code for Heima☆50Updated 2 months ago
- CorDA: Context-Oriented Decomposition Adaptation of Large Language Models for task-aware parameter-efficient fine-tuning(NeurIPS 2024)☆47Updated 6 months ago
- [EMNLP 2023, Main Conference] Sparse Low-rank Adaptation of Pre-trained Language Models☆79Updated last year
- A generalized framework for subspace tuning methods in parameter efficient fine-tuning.☆148Updated 3 weeks ago
- ☆16Updated 2 months ago
- [NeurIPS 2024] A Novel Rank-Based Metric for Evaluating Large Language Models☆50Updated last month
- [ICML 2024] Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibrati…☆40Updated last year
- Adapting LLaMA Decoder to Vision Transformer☆28Updated last year
- [ICLR 2025] Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models☆118Updated last week
- [NIPS2023]Implementation of Foundation Model is Efficient Multimodal Multitask Model Selector☆37Updated last year
- [ICLR 2025] When Attention Sink Emerges in Language Models: An Empirical View (Spotlight)☆99Updated last week
- [NeurIPS 2024] For paper Parameter Competition Balancing for Model Merging☆44Updated 9 months ago
- 🚀 LLaMA-MoE v2: Exploring Sparsity of LLaMA from Perspective of Mixture-of-Experts with Post-Training☆86Updated 7 months ago
- Doodling our way to AGI ✏️ 🖼️ 🧠☆72Updated last month
- Official implementation of "Traceable Evidence Enhanced Visual Grounded Reasoning: Evaluation and Methodology"☆41Updated last week
- ☆51Updated last week
- [ICLR 2025] Official implementation of paper "Dynamic Low-Rank Sparse Adaptation for Large Language Models".☆19Updated 4 months ago
- [MM2024, oral] "Self-Supervised Visual Preference Alignment" https://arxiv.org/abs/2404.10501☆56Updated 11 months ago