ZBox1005 / CoT-UQLinks
[arXiv 2025] "CoT-UQ: Improving Response-wise Uncertainty Quantification in LLMs with Chain-of-Thought"
☆14Updated 8 months ago
Alternatives and similar repositories for CoT-UQ
Users that are interested in CoT-UQ are comparing it to the libraries listed below
Sorting:
- Code for Heima☆58Updated 7 months ago
- Code for "Language Models Can Learn from Verbal Feedback Without Scalar Rewards"☆54Updated 2 months ago
- ☆142Updated 7 months ago
- JudgeLRM: Large Reasoning Models as a Judge☆40Updated this week
- [ACL'25 Oral] What Happened in LLMs Layers when Trained for Fast vs. Slow Thinking: A Gradient Perspective☆74Updated 5 months ago
- The code of RouterDC☆69Updated 7 months ago
- Discriminative Constrained Optimization for Reinforcing Large Reasoning Models☆46Updated last month
- [COLING'25] Exploring Concept Depth: How Large Language Models Acquire Knowledge at Different Layers?☆82Updated 10 months ago
- Official repository for paper: O1-Pruner: Length-Harmonizing Fine-Tuning for O1-Like Reasoning Pruning☆98Updated 9 months ago
- Official implementation of the NeurIPS 2025 paper "Soft Thinking: Unlocking the Reasoning Potential of LLMs in Continuous Concept Space"☆286Updated this week
- OpenVLThinker: An Early Exploration to Vision-Language Reasoning via Iterative Self-Improvement☆123Updated 4 months ago
- Code for paper "Unraveling Cross-Modality Knowledge Conflicts in Large Vision-Language Models."☆49Updated last year
- [ICCV 2025] Auto Interpretation Pipeline and many other functionalities for Multimodal SAE Analysis.☆166Updated 2 months ago
- Official repository for Montessori-Instruct: Generate Influential Training Data Tailored for Student Learning [ICLR 2025]☆50Updated 10 months ago
- Ruler: A Model-Agnostic Method to Control Generated Length for Large Language Models☆39Updated last year
- ☆70Updated last month
- ☆344Updated 4 months ago
- [ACL 2025 Findings] Implicit Reasoning in Transformers is Reasoning through Shortcuts☆17Updated 9 months ago
- Initialization using Update Approximation is a Silver Bullet for Extremely Efficient Low-Rank Fine-Tuning☆52Updated last month
- Enhancing Large Vision Language Models with Self-Training on Image Comprehension.☆70Updated last year
- Official Code for paper "Towards Efficient and Effective Unlearning of Large Language Models for Recommendation" (Frontiers of Computer S…☆38Updated last year
- [ICML'25] Our study systematically investigates massive values in LLMs' attention mechanisms. First, we observe massive values are concen…☆86Updated 5 months ago
- X-Reasoner: Towards Generalizable Reasoning Across Modalities and Domains☆49Updated 7 months ago
- AnchorAttention: Improved attention for LLMs long-context training☆213Updated 10 months ago
- [NeurIPS 2025] Thinkless: LLM Learns When to Think☆245Updated 2 months ago
- ☆51Updated 2 months ago
- [ACL 2025] A Generalizable and Purely Unsupervised Self-Training Framework☆70Updated 6 months ago
- [NeurIPS 2025] NoisyRollout: Reinforcing Visual Reasoning with Data Augmentation☆100Updated 2 months ago
- One-shot Entropy Minimization☆187Updated 6 months ago
- [ICML 2025] M-STAR (Multimodal Self-Evolving TrAining for Reasoning) Project. Diving into Self-Evolving Training for Multimodal Reasoning☆69Updated 5 months ago