waltonfuture / Diff-eRankLinks
[NeurIPS 2024] A Novel Rank-Based Metric for Evaluating Large Language Models
☆49Updated last month
Alternatives and similar repositories for Diff-eRank
Users that are interested in Diff-eRank are comparing it to the libraries listed below
Sorting:
- ☆113Updated 4 months ago
- CoT-Valve: Length-Compressible Chain-of-Thought Tuning☆76Updated 5 months ago
- ☆132Updated last month
- A curated list of awesome LLM Inference-Time Self-Improvement (ITSI, pronounced "itsy") papers from our recent survey: A Survey on Large …☆84Updated 6 months ago
- 📖 This is a repository for organizing papers, codes, and other resources related to Latent Reasoning.☆86Updated this week
- ☆122Updated last month
- [ICLR 2025] SuperCorrect: Advancing Small LLM Reasoning with Thought Template Distillation and Self-Correction☆74Updated 3 months ago
- The official repository of paper "AdaR1: From Long-Cot to Hybrid-CoT via Bi-Level Adaptive Reasoning Optimization"☆18Updated 2 months ago
- [NeurIPS-2024] 📈 Scaling Laws with Vocabulary: Larger Models Deserve Larger Vocabularies https://arxiv.org/abs/2407.13623☆86Updated 9 months ago
- One-shot Entropy Minimization☆165Updated last month
- A Sober Look at Language Model Reasoning☆75Updated last month
- ☆318Updated last month
- [ICLR 2025] When Attention Sink Emerges in Language Models: An Empirical View (Spotlight)☆96Updated last week
- [ICLR 2025] Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models☆116Updated last week
- RM-R1: Unleashing the Reasoning Potential of Reward Models☆113Updated 2 weeks ago
- 🚀 LLaMA-MoE v2: Exploring Sparsity of LLaMA from Perspective of Mixture-of-Experts with Post-Training☆86Updated 7 months ago
- SLED: Self Logits Evolution Decoding for Improving Factuality in Large Language Model https://arxiv.org/pdf/2411.02433☆27Updated 7 months ago
- Official Implementation for EMNLP 2024 (main) "AgentReview: Exploring Academic Peer Review with LLM Agent."☆75Updated 8 months ago
- ☆47Updated 4 months ago
- [ICML'25] Our study systematically investigates massive values in LLMs' attention mechanisms. First, we observe massive values are concen…☆74Updated 3 weeks ago
- ☆90Updated 2 months ago
- [ICML 2024] Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibrati…☆40Updated last year
- Official codebase for "GenPRM: Scaling Test-Time Compute of Process Reward Models via Generative Reasoning".☆78Updated last month
- [ACL'25 Oral] What Happened in LLMs Layers when Trained for Fast vs. Slow Thinking: A Gradient Perspective☆70Updated 3 weeks ago
- ☆64Updated 3 weeks ago
- Large Language Models Can Self-Improve in Long-context Reasoning☆71Updated 7 months ago
- ☆19Updated 2 months ago
- End-to-End Reinforcement Learning for Multi-Turn Tool-Integrated Reasoning☆113Updated last week
- Official code of *Virgo: A Preliminary Exploration on Reproducing o1-like MLLM*☆105Updated last month
- A comrephensive collection of learning from rewards in the post-training and test-time scaling of LLMs, with a focus on both reward model…☆50Updated last month