linlu-qiu / lm-inductive-reasoningLinks
☆33Updated last year
Alternatives and similar repositories for lm-inductive-reasoning
Users that are interested in lm-inductive-reasoning are comparing it to the libraries listed below
Sorting:
- ☆57Updated 5 months ago
- Official code for paper Understanding the Reasoning Ability of Language Models From the Perspective of Reasoning Paths Aggregation☆20Updated last year
- Self-Supervised Alignment with Mutual Information☆21Updated last year
- [ICLR 2023] Code for our paper "Selective Annotation Makes Language Models Better Few-Shot Learners"☆111Updated 2 years ago
- The Unreliability of Explanations in Few-shot Prompting for Textual Reasoning (NeurIPS 2022)☆16Updated 2 years ago
- Easy-to-Hard Generalization: Scalable Alignment Beyond Human Supervision☆125Updated last year
- ☆85Updated last year
- ☆103Updated last year
- [ICLR 2025] Unintentional Unalignment: Likelihood Displacement in Direct Preference Optimization☆31Updated 9 months ago
- ☆129Updated last year
- The accompanying code for "Transformer Feed-Forward Layers Are Key-Value Memories". Mor Geva, Roei Schuster, Jonathan Berant, and Omer Le…☆99Updated 4 years ago
- [EMNLP-2022 Findings] Code for paper “ProGen: Progressive Zero-shot Dataset Generation via In-context Feedback”.☆27Updated 2 years ago
- A framework to train language models to learn invariant representations.☆14Updated 3 years ago
- ☆36Updated last year
- Supporting code for ReCEval paper☆30Updated last year
- Language modeling via stochastic processes. Oral @ ICLR 2022.☆138Updated 2 years ago
- ☆113Updated 3 years ago
- Offical code of the paper Large Language Models Are Implicitly Topic Models: Explaining and Finding Good Demonstrations for In-Context Le…☆75Updated last year
- ☆35Updated 2 years ago
- ☆54Updated 2 years ago
- Personalized Soups: Personalized Large Language Model Alignment via Post-hoc Parameter Merging☆111Updated 2 years ago
- Synthetic question-answering dataset to formally analyze the chain-of-thought output of large language models on a reasoning task.☆151Updated 2 months ago
- ☆104Updated last year
- ☆27Updated 2 years ago
- Source codes for "Preference-grounded Token-level Guidance for Language Model Fine-tuning" (NeurIPS 2023).☆16Updated 10 months ago
- Code for our paper: "GrIPS: Gradient-free, Edit-based Instruction Search for Prompting Large Language Models"☆57Updated 2 years ago
- Code for RL4F: Generating Natural Language Feedback with Reinforcement Learning for Repairing Model Outputs. ACL 2023.☆64Updated 11 months ago
- [ICML 2023] Code for our paper “Compositional Exemplars for In-context Learning”.☆103Updated 2 years ago
- Code for Paper (Preserving Diversity in Supervised Fine-tuning of Large Language Models)☆47Updated 6 months ago
- Directional Preference Alignment☆57Updated last year