WANGXinyiLinda / LM_random_walkLinks
Official code for paper Understanding the Reasoning Ability of Language Models From the Perspective of Reasoning Paths Aggregation
☆20Updated last year
Alternatives and similar repositories for LM_random_walk
Users that are interested in LM_random_walk are comparing it to the libraries listed below
Sorting:
- [NAACL'25 Oral] Steering Knowledge Selection Behaviours in LLMs via SAE-Based Representation Engineering☆63Updated 10 months ago
- ☆44Updated last year
- Source codes for "Preference-grounded Token-level Guidance for Language Model Fine-tuning" (NeurIPS 2023).☆16Updated 8 months ago
- [EMNLP 2023] MQuAKE: Assessing Knowledge Editing in Language Models via Multi-Hop Questions☆115Updated last year
- LogicBench is a natural language question-answering dataset consisting of 25 different reasoning patterns spanning over propositional, fi…☆29Updated last year
- ☆43Updated 6 months ago
- ☆29Updated last year
- ☆97Updated last year
- ☆68Updated last year
- ☆57Updated 4 months ago
- A Mechanistic Understanding of Alignment Algorithms: A Case Study on DPO and Toxicity.☆78Updated 6 months ago
- Easy-to-Hard Generalization: Scalable Alignment Beyond Human Supervision☆123Updated last year
- The code of “Improving Weak-to-Strong Generalization with Scalable Oversight and Ensemble Learning”☆17Updated last year
- Augmenting Statistical Models with Natural Language Parameters☆27Updated last year
- ☆15Updated last year
- Code for the ICML 2024 paper "Rewards-in-Context: Multi-objective Alignment of Foundation Models with Dynamic Preference Adjustment"☆74Updated 3 months ago
- ☆33Updated last year
- GenRM-CoT: Data release for verification rationales☆65Updated 11 months ago
- Code and data used in the paper: "Training on Incorrect Synthetic Data via RL Scales LLM Math Reasoning Eight-Fold"☆30Updated last year
- Personalized Soups: Personalized Large Language Model Alignment via Post-hoc Parameter Merging☆108Updated last year
- The Unreliability of Explanations in Few-shot Prompting for Textual Reasoning (NeurIPS 2022)☆16Updated 2 years ago
- ☆101Updated last year
- [ICLR 2025] Unintentional Unalignment: Likelihood Displacement in Direct Preference Optimization☆31Updated 8 months ago
- ☆17Updated last year
- AbstainQA, ACL 2024☆28Updated 11 months ago
- ☆28Updated last year
- Search, Verify and Feedback: Towards Next Generation Post-training Paradigm of Foundation Models via Verifier Engineering☆61Updated 9 months ago
- [COLM'24] "Deductive Beam Search: Decoding Deducible Rationale for Chain-of-Thought Reasoning"☆21Updated last year
- ☆55Updated 2 years ago
- Offical code of the paper Large Language Models Are Implicitly Topic Models: Explaining and Finding Good Demonstrations for In-Context Le…☆75Updated last year