xinzhel / LLM-SearchLinks
Survey on LLM Inference via Search (TMLR 2025)
☆10Updated 3 months ago
Alternatives and similar repositories for LLM-Search
Users that are interested in LLM-Search are comparing it to the libraries listed below
Sorting:
- CoT-Valve: Length-Compressible Chain-of-Thought Tuning☆82Updated 5 months ago
- ☆156Updated 2 months ago
- Code repo for "Harnessing Negative Signals: Reinforcement Distillation from Teacher Data for LLM Reasoning"☆27Updated 2 weeks ago
- [ICLR 2025] Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models☆121Updated last month
- [ICML'25] Our study systematically investigates massive values in LLMs' attention mechanisms. First, we observe massive values are concen…☆77Updated last month
- [arXiv] Do Not Let Low-Probability Tokens Over-Dominate in RL for LLMs☆36Updated 2 months ago
- Paper List of Inference/Test Time Scaling/Computing☆289Updated last month
- [ICLR 2025] When Attention Sink Emerges in Language Models: An Empirical View (Spotlight)☆107Updated last month
- [ICML 2024] Official code for the paper "Revisiting Zeroth-Order Optimization for Memory-Efficient LLM Fine-Tuning: A Benchmark ".☆109Updated last month
- Official PyTorch code for ICLR 2025 paper "Gnothi Seauton: Empowering Faithful Self-Interpretability in Black-Box Models"☆20Updated 5 months ago
- [arXiv 2025] Efficient Reasoning Models: A Survey☆247Updated 3 weeks ago
- ☆31Updated 3 months ago
- The repository of the paper "REEF: Representation Encoding Fingerprints for Large Language Models," aims to protect the IP of open-source…☆59Updated 6 months ago
- [ICML 2024] Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibrati…☆40Updated last year
- Implementation of the MATRIX framework (ICML 2024)☆58Updated last year
- Accepted LLM Papers in NeurIPS 2024☆37Updated 9 months ago
- Preprint: Asymmetry in Low-Rank Adapters of Foundation Models☆35Updated last year
- 🚀 LLaMA-MoE v2: Exploring Sparsity of LLaMA from Perspective of Mixture-of-Experts with Post-Training☆86Updated 8 months ago
- Official Repository of "Learning what reinforcement learning can't"☆54Updated last week
- PoC for "SpecReason: Fast and Accurate Inference-Time Compute via Speculative Reasoning" [arXiv '25]☆45Updated last month
- Official Repository of LatentSeek☆56Updated 2 months ago
- A comprehensive collection of process reward models.☆99Updated 3 weeks ago
- "what, how, where, and how well? a survey on test-time scaling in large language models" repository☆57Updated this week
- A Sober Look at Language Model Reasoning☆81Updated last month
- Must-read papers and blogs about parametric knowledge mechanism in LLMs.☆21Updated 3 months ago
- Official PyTorch implementation of the paper "dLLM-Cache: Accelerating Diffusion Large Language Models with Adaptive Caching" (dLLM-Cache…☆134Updated this week
- Official Pytorch Implementation of Our Paper Accepted at ICLR 2024-- Dynamic Sparse No Training: Training-Free Fine-tuning for Sparse LLM…☆49Updated last year
- Optimizing Anytime Reasoning via Budget Relative Policy Optimization☆43Updated 3 weeks ago
- [NeurIPS 2024] The official implementation of ZipCache: Accurate and Efficient KV Cache Quantization with Salient Token Identification☆23Updated 4 months ago
- [EMNLP 2024 Findings🔥] Official implementation of ": LOOK-M: Look-Once Optimization in KV Cache for Efficient Multimodal Long-Context In…☆98Updated 9 months ago