SalesforceAIResearch / MobileAIBenchLinks
☆24Updated 3 months ago
Alternatives and similar repositories for MobileAIBench
Users that are interested in MobileAIBench are comparing it to the libraries listed below
Sorting:
- Co-LLM: Learning to Decode Collaboratively with Multiple Language Models☆126Updated last year
- ☆209Updated 2 years ago
- The official repo for "LLoCo: Learning Long Contexts Offline"☆118Updated last year
- Pytorch implementation for "Compressed Context Memory For Online Language Model Interaction" (ICLR'24)☆62Updated last year
- [NAACL 2025] A Closer Look into Mixture-of-Experts in Large Language Models☆60Updated last year
- Code for PHATGOOSE introduced in "Learning to Route Among Specialized Experts for Zero-Shot Generalization"☆91Updated last year
- Repository for NPHardEval, a quantified-dynamic benchmark of LLMs☆63Updated last year
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks (EMNLP'24)☆147Updated last year
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆144Updated 2 years ago
- [NeurIPS 2024] Goldfish Loss: Mitigating Memorization in Generative LLMs☆94Updated last year
- ☆273Updated 2 years ago
- [TMLR 2026] When Attention Collapses: How Degenerate Layers in LLMs Enable Smaller, Stronger Models☆122Updated last year
- ☆74Updated last year
- Fast and Robust Early-Exiting Framework for Autoregressive Language Models with Synchronized Parallel Decoding (EMNLP 2023 Long)☆65Updated last year
- Code and Data for "Long-context LLMs Struggle with Long In-context Learning" [TMLR2025]☆111Updated 11 months ago
- [NeurIPS'23] Aging with GRACE: Lifelong Model Editing with Discrete Key-Value Adaptors☆83Updated last year
- Code Implementation, Evaluations, Documentation, Links and Resources for Min P paper☆46Updated 5 months ago
- [ICML 2025] From Low Rank Gradient Subspace Stabilization to Low-Rank Weights: Observations, Theories and Applications☆52Updated 3 months ago
- PASTA: Post-hoc Attention Steering for LLMs☆134Updated last year
- Test-time-training on nearest neighbors for large language models☆49Updated last year
- ☆52Updated 10 months ago
- [ICLR 2023] "Sparse MoE as the New Dropout: Scaling Dense and Self-Slimmable Transformers" by Tianlong Chen*, Zhenyu Zhang*, Ajay Jaiswal…☆56Updated 2 years ago
- Replicating O1 inference-time scaling laws☆93Updated last year
- This is official project in our paper: Is Bigger and Deeper Always Better? Probing LLaMA Across Scales and Layers☆31Updated 2 years ago
- ☆41Updated 2 years ago
- Unofficial Implementation of Chain-of-Thought Reasoning Without Prompting☆35Updated last year
- Code associated with Tuning Language Models by Proxy (Liu et al., 2024)☆127Updated last year
- Official github repo for the paper "Compression Represents Intelligence Linearly" [COLM 2024]☆147Updated last year
- [NeurIPS 2024] Knowledge Circuits in Pretrained Transformers☆163Updated 2 months ago
- ☆56Updated 10 months ago