☆62May 13, 2025Updated 9 months ago
Alternatives and similar repositories for PPE
Users that are interested in PPE are comparing it to the libraries listed below
Sorting:
- ☆47Mar 25, 2025Updated 11 months ago
- Official repository for ACL 2025 paper "ProcessBench: Identifying Process Errors in Mathematical Reasoning"☆184May 20, 2025Updated 9 months ago
- ☆13Dec 17, 2025Updated 2 months ago
- LongAttn :Selecting Long-context Training Data via Token-level Attention☆15Jul 16, 2025Updated 7 months ago
- ☆21Jan 4, 2026Updated last month
- ☆34Nov 26, 2025Updated 3 months ago
- The repository contains code for Adaptive Data Optimization☆32Dec 9, 2024Updated last year
- RM-R1: Unleashing the Reasoning Potential of Reward Models☆159Jun 26, 2025Updated 8 months ago
- ☆17Mar 3, 2025Updated last year
- Benchmarking Complex Instruction-Following with Multiple Constraints Composition (NeurIPS 2024 Datasets and Benchmarks Track)☆102Feb 20, 2025Updated last year
- Tasks and tutorials using Graphore's IPU with Hugging Face. Originally at https://github.com/gradient-ai/Graphcore-HuggingFace☆16Mar 12, 2024Updated last year
- ☆39Feb 11, 2026Updated 2 weeks ago
- The code for paper "EPO: Entropy-regularized Policy Optimization for LLM Agents Reinforcement Learning"☆37Oct 1, 2025Updated 5 months ago
- A Recipe for Building LLM Reasoners to Solve Complex Instructions☆29Oct 9, 2025Updated 4 months ago
- Official repository for "BLEUBERI: BLEU is a surprisingly effective reward for instruction following"☆31Jun 5, 2025Updated 8 months ago
- A simple implementation of ReasonGenRM.☆19Apr 21, 2025Updated 10 months ago
- ☆65Feb 12, 2026Updated 2 weeks ago
- Arena-Hard-Auto: An automatic LLM benchmark.☆1,003Jun 21, 2025Updated 8 months ago
- Systematic evaluation framework that automatically rates overthinking behavior in large language models.☆96May 16, 2025Updated 9 months ago
- Recipes to train reward model for RLHF.☆1,517Apr 24, 2025Updated 10 months ago
- RewardBench: the first evaluation tool for reward models.☆696Feb 16, 2026Updated 2 weeks ago
- Self-Supervised Alignment with Mutual Information☆20May 24, 2024Updated last year
- ☆31Sep 12, 2025Updated 5 months ago
- Learning to route instances for Human vs AI Feedback (ACL Main '25)☆27Jul 23, 2025Updated 7 months ago
- Your personal ArXiv Feed☆23Dec 18, 2024Updated last year
- Code for our EMNLP-2023 paper: "Active Instruction Tuning: Improving Cross-Task Generalization by Training on Prompt Sensitive Tasks"☆25Nov 16, 2023Updated 2 years ago
- [ACL 2024 Findings] CriticBench: Benchmarking LLMs for Critique-Correct Reasoning☆30Mar 5, 2024Updated last year
- Code base for internal reward models and PPO training☆24Oct 1, 2023Updated 2 years ago
- ☆53Feb 11, 2025Updated last year
- [ICLR 2025] SWIFT: On-the-Fly Self-Speculative Decoding for LLM Inference Acceleration☆62Feb 21, 2025Updated last year
- ☆47Oct 2, 2025Updated 5 months ago
- NExT-GPT: Any-to-Any Multimodal Large Language Model☆20Nov 3, 2024Updated last year
- Code for Evolving Language Models without Labels: Majority Drives Selection, Novelty Promotes Variation (EVOL-RL).☆48Oct 16, 2025Updated 4 months ago
- ☆59Aug 22, 2024Updated last year
- Super-Efficient RLHF Training of LLMs with Parameter Reallocation☆331Apr 24, 2025Updated 10 months ago
- ☆43Aug 31, 2025Updated 6 months ago
- Code for the paper "VinePPO: Unlocking RL Potential For LLM Reasoning Through Refined Credit Assignment"☆186May 25, 2025Updated 9 months ago
- ☆109Dec 10, 2025Updated 2 months ago
- PyTorch implementation of the Region Mutual Information Loss for Semantic Segmentation.☆26Oct 26, 2023Updated 2 years ago