☆131Oct 1, 2024Updated last year
Alternatives and similar repositories for demonstrated-feedback
Users that are interested in demonstrated-feedback are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- Official repository for ORPO☆473May 31, 2024Updated last year
- Code for the paper Task Agnostic Morphology Evolution.☆20May 25, 2021Updated 4 years ago
- Code accompanying the paper "Noise Contrastive Alignment of Language Models with Explicit Rewards" (NeurIPS 2024)☆58Nov 8, 2024Updated last year
- ☆16Jul 23, 2024Updated last year
- ☆23Sep 19, 2024Updated last year
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment☆62Aug 30, 2024Updated last year
- ☆15Oct 4, 2024Updated last year
- Self-Supervised Alignment with Mutual Information☆20May 24, 2024Updated last year
- BERT score for text generation☆12Jan 15, 2025Updated last year
- ☆21Jul 21, 2025Updated 8 months ago
- RS-IMLE☆44Dec 7, 2024Updated last year
- ☆12Jun 20, 2023Updated 2 years ago
- Official code for "MAmmoTH2: Scaling Instructions from the Web" [NeurIPS 2024]☆149Oct 27, 2024Updated last year
- Aligning Agentic World Models via Knowledgeable Experience Learning☆32Jan 25, 2026Updated 2 months ago
- Code and data from the paper 'Human Feedback is not Gold Standard'☆20Mar 6, 2026Updated 2 weeks ago
- ☆58Sep 2, 2024Updated last year
- Code for RATIONALYST: Pre-training Process-Supervision for Improving Reasoning https://arxiv.org/pdf/2410.01044☆35Oct 3, 2024Updated last year
- The official implementation of Self-Play Fine-Tuning (SPIN)☆1,234May 8, 2024Updated last year
- Self-Alignment with Principle-Following Reward Models☆170Sep 18, 2025Updated 6 months ago
- A library with extensible implementations of DPO, KTO, PPO, ORPO, and other human-aware loss functions (HALOs).☆907Sep 30, 2025Updated 5 months ago
- Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Fl…☆78Aug 17, 2024Updated last year
- B-STAR: Monitoring and Balancing Exploration and Exploitation in Self-Taught Reasoners☆86May 21, 2025Updated 10 months ago
- Scalable toolkit for efficient model alignment☆850Oct 6, 2025Updated 5 months ago
- Layer-Condensed KV cache w/ 10 times larger batch size, fewer params and less computation. Dramatic speed up with better task performance…☆157Apr 7, 2025Updated 11 months ago
- ☆18Sep 12, 2021Updated 4 years ago
- The official evaluation suite and dynamic data release for MixEval.☆256Nov 10, 2024Updated last year
- This is the oficial repository for "Safer-Instruct: Aligning Language Models with Automated Preference Data"☆17Feb 22, 2024Updated 2 years ago
- Original reference implementation of "Analyzing the Internals of Neural Radiance Fields"☆11Apr 10, 2024Updated last year
- Learning from preferences is a common paradigm for fine-tuning language models. Yet, many algorithmic design decisions come into play. Ou…☆32Apr 20, 2024Updated last year
- A RAG that can scale 🧑🏻💻☆11May 28, 2024Updated last year
- Dateset Reset Policy Optimization☆31Apr 12, 2024Updated last year
- Code for Zero-Shot Tokenizer Transfer☆143Jan 14, 2025Updated last year
- [𝐄𝐌𝐍𝐋𝐏 𝐅𝐢𝐧𝐝𝐢𝐧𝐠𝐬 𝟐𝟎𝟐𝟒 & 𝐀𝐂𝐋 𝟐𝟎𝟐𝟒 𝐍𝐋𝐑𝐒𝐄 𝐎𝐫𝐚𝐥] 𝘌𝘯𝘩𝘢𝘯𝘤𝘪𝘯𝘨 𝘔𝘢𝘵𝘩𝘦𝘮𝘢𝘵𝘪𝘤𝘢𝘭 𝘙𝘦𝘢𝘴 𝘰𝘯𝘪𝘯…☆51May 4, 2024Updated last year
- Pretraining summarization models using a corpus of nonsense☆13Sep 28, 2021Updated 4 years ago
- Code for NeurIPS'24 paper 'Grokked Transformers are Implicit Reasoners: A Mechanistic Journey to the Edge of Generalization'☆234Jul 19, 2025Updated 8 months ago
- [ACL 2025] An inference-time decoding strategy with adaptive foresight sampling☆108May 18, 2025Updated 10 months ago
- A recipe for online RLHF and online iterative DPO.☆545Dec 28, 2024Updated last year
- [NeurIPS 2024] SimPO: Simple Preference Optimization with a Reference-Free Reward☆948Feb 16, 2025Updated last year
- ☆327Jul 25, 2024Updated last year