Moocember / Optimization-by-PROmptingLinks
☆78Updated last year
Alternatives and similar repositories for Optimization-by-PROmpting
Users that are interested in Optimization-by-PROmpting are comparing it to the libraries listed below
Sorting:
- A dataset of LLM-generated chain-of-thought steps annotated with mistake location.☆81Updated 11 months ago
- ☆124Updated 9 months ago
- Scripts for generating synthetic finetuning data for reducing sycophancy.☆113Updated last year
- Meta-CoT: Generalizable Chain-of-Thought Prompting in Mixed-task Scenarios with Large Language Models☆97Updated last year
- ☆121Updated last year
- Astraios: Parameter-Efficient Instruction Tuning Code Language Models☆58Updated last year
- ☆150Updated last year
- Challenge LLMs to Reason About Reasoning: A Benchmark to Unveil Cognitive Depth in LLMs☆49Updated last year
- CodeUltraFeedback: aligning large language models to coding preferences☆71Updated last year
- Open Implementations of LLM Analyses☆105Updated 9 months ago
- ☆119Updated 11 months ago
- Repository for NPHardEval, a quantified-dynamic benchmark of LLMs☆56Updated last year
- Scalable Meta-Evaluation of LLMs as Evaluators☆42Updated last year
- augmented LLM with self reflection☆129Updated last year
- Official repo of Respond-and-Respond: data, code, and evaluation☆103Updated 11 months ago
- ☆98Updated last year
- Code and data accompanying our paper on arXiv "Faithful Chain-of-Thought Reasoning".☆161Updated last year
- Plug in and play implementation of " Textbooks Are All You Need", ready for training, inference, and dataset generation☆76Updated last year
- Codebase accompanying the Summary of a Haystack paper.☆79Updated 9 months ago
- Evaluating LLMs with CommonGen-Lite☆90Updated last year
- Code for PHATGOOSE introduced in "Learning to Route Among Specialized Experts for Zero-Shot Generalization"☆86Updated last year
- Reformatted Alignment☆113Updated 9 months ago
- Q-Probe: A Lightweight Approach to Reward Maximization for Language Models☆41Updated last year
- Replicating O1 inference-time scaling laws☆87Updated 7 months ago
- [ICML 2025] Flow of Reasoning: Training LLMs for Divergent Reasoning with Minimal Examples☆100Updated last month
- NeurIPS 2024 tutorial on LLM Inference☆45Updated 7 months ago
- Minimal implementation of the Self-Play Fine-Tuning Converts Weak Language Models to Strong Language Models paper (ArXiv 20232401.01335)☆28Updated last year
- FuseAI Project☆87Updated 5 months ago
- ☆133Updated last year
- ☆183Updated 5 months ago