lukasberglund / reversal_curseLinks
☆288Updated last year
Alternatives and similar repositories for reversal_curse
Users that are interested in reversal_curse are comparing it to the libraries listed below
Sorting:
- Code and data accompanying our paper on arXiv "Faithful Chain-of-Thought Reasoning".☆159Updated last year
- Mass-editing thousands of facts into a transformer memory (ICLR 2023)☆492Updated last year
- Simple next-token-prediction for RLHF☆226Updated last year
- ☆131Updated 6 months ago
- The dataset and code for paper: TheoremQA: A Theorem-driven Question Answering dataset☆157Updated last year
- ☆133Updated last year
- [EMNLP 2023] The CoT Collection: Improving Zero-shot and Few-shot Learning of Language Models via Chain-of-Thought Fine-Tuning☆241Updated last year
- Inference-Time Intervention: Eliciting Truthful Answers from a Language Model☆526Updated 4 months ago
- Evaluating LLMs with fewer examples☆156Updated last year
- PASTA: Post-hoc Attention Steering for LLMs☆117Updated 6 months ago
- ☆149Updated last year
- Function Vectors in Large Language Models (ICLR 2024)☆167Updated last month
- Official github repo for the paper "Compression Represents Intelligence Linearly" [COLM 2024]☆135Updated 8 months ago
- Code and data for "Lost in the Middle: How Language Models Use Long Contexts"☆344Updated last year
- Code for NeurIPS'24 paper 'Grokked Transformers are Implicit Reasoners: A Mechanistic Journey to the Edge of Generalization'☆193Updated 6 months ago
- Self-Alignment with Principle-Following Reward Models☆161Updated 3 weeks ago
- Inspecting and Editing Knowledge Representations in Language Models☆116Updated last year
- This is work done by the Oxen.ai Community, trying to reproduce the Self-Rewarding Language Model paper from MetaAI.☆128Updated 6 months ago
- Benchmarking LLMs with Challenging Tasks from Real Users☆223Updated 6 months ago
- RuLES: a benchmark for evaluating rule-following in language models☆224Updated 3 months ago
- ☆174Updated last month
- Learning to Compress Prompts with Gist Tokens - https://arxiv.org/abs/2304.08467☆285Updated 3 months ago
- [NeurIPS 2024] Knowledge Circuits in Pretrained Transformers☆148Updated 3 months ago
- Challenging BIG-Bench Tasks and Whether Chain-of-Thought Can Solve Them☆493Updated 11 months ago
- [EMNLP 2023] Adapting Language Models to Compress Long Contexts☆305Updated 8 months ago
- Tools for understanding how transformer predictions are built layer-by-layer☆493Updated last year
- ☆120Updated 8 months ago
- Functional Benchmarks and the Reasoning Gap☆86Updated 8 months ago
- Code to reproduce "Transformers Can Do Arithmetic with the Right Embeddings", McLeish et al (NeurIPS 2024)☆189Updated last year
- TART: A plug-and-play Transformer module for task-agnostic reasoning☆196Updated last year