DeqingFu / transformers-icl-second-orderLinks
Official repository for our paper, Transformers Learn Higher-Order Optimization Methods for In-Context Learning: A Study with Linear Models.
☆16Updated 6 months ago
Alternatives and similar repositories for transformers-icl-second-order
Users that are interested in transformers-icl-second-order are comparing it to the libraries listed below
Sorting:
- ☆40Updated last year
- Rewarded soups official implementation☆58Updated last year
- official code for paper Probing the Decision Boundaries of In-context Learning in Large Language Models. https://arxiv.org/abs/2406.11233…☆18Updated 9 months ago
- What Makes a Reward Model a Good Teacher? An Optimization Perspective☆31Updated last month
- official implementation of ICLR'2025 paper: Rethinking Bradley-Terry Models in Preference-based Reward Modeling: Foundations, Theory, and…☆58Updated 2 months ago
- ☆18Updated last year
- Align your LM to express calibrated verbal statements of confidence in its long-form generations.☆25Updated last year
- A Mechanistic Understanding of Alignment Algorithms: A Case Study on DPO and Toxicity.☆72Updated 2 months ago
- Directional Preference Alignment☆56Updated 8 months ago
- Code for Paper (Policy Optimization in RLHF: The Impact of Out-of-preference Data)☆28Updated last year
- ☆29Updated last year
- ☆32Updated last year
- Code release for "Debating with More Persuasive LLMs Leads to More Truthful Answers"