benpry / why-think-step-by-step
Code and data for the paper "Why think step by step? Reasoning emerges from the locality of experience"
☆56Updated last year
Alternatives and similar repositories for why-think-step-by-step:
Users that are interested in why-think-step-by-step are comparing it to the libraries listed below
- Learning to Retrieve by Trying - Source code for Grounding by Trying: LLMs with Reinforcement Learning-Enhanced Retrieval☆29Updated 3 months ago
- Implementation of the paper: "AssistantBench: Can Web Agents Solve Realistic and Time-Consuming Tasks?"☆47Updated last month
- Evaluating LLMs with CommonGen-Lite☆88Updated 10 months ago
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment☆53Updated 5 months ago
- Functional Benchmarks and the Reasoning Gap☆82Updated 3 months ago
- Code for EMNLP 2024 paper "Learn Beyond The Answer: Training Language Models with Reflection for Mathematical Reasoning"☆50Updated 3 months ago
- An automated tool for discovering insights from research papaer corpora☆136Updated 7 months ago
- ☆59Updated 9 months ago
- A repository for research on medium sized language models.☆76Updated 8 months ago
- Code for RATIONALYST: Pre-training Process-Supervision for Improving Reasoning https://arxiv.org/pdf/2410.01044☆32Updated 3 months ago
- LLMs as Collaboratively Edited Knowledge Bases☆43Updated 11 months ago
- OpenCoconut implements a latent reasoning paradigm where we generate thoughts before decoding.☆158Updated 2 weeks ago
- How to create rational LLM-based agents? Using game-theoretic workflows!☆47Updated last month
- LLM reads a paper and produce a working prototype☆48Updated last month
- ☆30Updated 4 months ago
- Code for our paper PAPILLON: PrivAcy Preservation from Internet-based and Local Language MOdel ENsembles☆20Updated last month
- Script for processing OpenAI's PRM800K process supervision dataset into an Alpaca-style instruction-response format☆27Updated last year
- LILO: Library Induction with Language Observations☆82Updated 5 months ago
- The code implementation of MAGDi: Structured Distillation of Multi-Agent Interaction Graphs Improves Reasoning in Smaller Language Models…☆31Updated 11 months ago
- Testing paligemma2 finetuning on reasoning dataset☆18Updated last month
- ☆98Updated last week
- Score LLM pretraining data with classifiers☆54Updated last year
- NeurIPS 2023 - Cappy: Outperforming and Boosting Large Multi-Task LMs with a Small Scorer☆40Updated 10 months ago
- Codebase accompanying the Summary of a Haystack paper.☆74Updated 4 months ago
- ☆48Updated 2 months ago
- Minimal implementation of the Self-Play Fine-Tuning Converts Weak Language Models to Strong Language Models paper (ArXiv 20232401.01335)☆29Updated 10 months ago
- ☆31Updated 7 months ago
- Q-Probe: A Lightweight Approach to Reward Maximization for Language Models☆40Updated 7 months ago