HKUDS / SepLLMLinks
[ICML 2025] "SepLLM: Accelerate Large Language Models by Compressing One Segment into One Separator"
☆75Updated this week
Alternatives and similar repositories for SepLLM
Users that are interested in SepLLM are comparing it to the libraries listed below
Sorting:
- Code for "Your Mixture-of-Experts LLM Is Secretly an Embedding Model For Free"☆74Updated 8 months ago
- CoT-Valve: Length-Compressible Chain-of-Thought Tuning☆73Updated 4 months ago
- [ACL'25] We propose a novel fine-tuning method, Separate Memory and Reasoning, which combines prompt tuning with LoRA.☆65Updated last month
- ☆22Updated 11 months ago
- Repo for "Z1: Efficient Test-time Scaling with Code"☆61Updated 2 months ago
- TokenSkip: Controllable Chain-of-Thought Compression in LLMs☆156Updated 3 months ago
- ☆24Updated 3 months ago
- ☆109Updated 3 months ago
- ☆29Updated 2 months ago
- [ICML 2025] Teaching Language Models to Critique via Reinforcement Learning☆99Updated last month
- ☆104Updated 3 weeks ago
- ☆116Updated 3 weeks ago
- A Sober Look at Language Model Reasoning☆74Updated last week
- Easy control for Key-Value Constrained Generative LLM Inference(https://arxiv.org/abs/2402.06262)☆63Updated last year
- ☆147Updated 9 months ago
- [ICML 2025] Reward-guided Speculative Decoding (RSD) for efficiency and effectiveness.☆32Updated last month
- Fira: Can We Achieve Full-rank Training of LLMs Under Low-rank Constraint?☆112Updated 8 months ago
- ☆37Updated 8 months ago
- "what, how, where, and how well? a survey on test-time scaling in large language models" repository☆45Updated this week
- ACL'2025: SoftCoT: Soft Chain-of-Thought for Efficient Reasoning with LLMs. and preprint: SoftCoT++: Test-Time Scaling with Soft Chain-of…☆28Updated 3 weeks ago
- [ACL 2024] Not All Experts are Equal: Efficient Expert Pruning and Skipping for Mixture-of-Experts Large Language Models☆92Updated last year
- HelloBench: Evaluating Long Text Generation Capabilities of Large Language Models☆45Updated 7 months ago
- RM-R1: Unleashing the Reasoning Potential of Reward Models☆108Updated 3 weeks ago
- [ICLR 2024 Spotlight] Code for the paper "Merge, Then Compress: Demystify Efficient SMoE with Hints from Its Routing Policy"☆85Updated this week
- Official codebase for "GenPRM: Scaling Test-Time Compute of Process Reward Models via Generative Reasoning".☆75Updated 3 weeks ago
- ☆58Updated this week
- AnchorAttention: Improved attention for LLMs long-context training☆208Updated 5 months ago
- [ICLR 2025 Workshop] "Landscape of Thoughts: Visualizing the Reasoning Process of Large Language Models"☆25Updated last week
- [ICML'25] Our study systematically investigates massive values in LLMs' attention mechanisms. First, we observe massive values are concen…☆73Updated this week
- SIFT: Grounding LLM Reasoning in Contexts via Stickers☆56Updated 3 months ago