ArcherCodeR is an open-source initiative enhancing code reasoning in large language models through scalable, rule-governed reinforcement learning.
☆44Aug 6, 2025Updated 6 months ago
Alternatives and similar repositories for ArcherCodeR
Users that are interested in ArcherCodeR are comparing it to the libraries listed below
Sorting:
- A scalable automated alignment method for large language models. Resources for "Aligning Large Language Models via Self-Steering Optimiza…☆20Nov 21, 2024Updated last year
- [ICML 2025] Beyond Bradley-Terry Models: A General Preference Model for Language Model Alignment (https://arxiv.org/abs/2410.02197)☆39Sep 8, 2025Updated 5 months ago
- [AAAI 2025]Automatically Generating Numerous Context-Driven SFT Data for LLMs across Diverse Granularity☆26Mar 17, 2025Updated 11 months ago
- Code for "Seeking Neural Nuggets: Knowledge Transfer in Large Language Models from a Parametric Perspective"☆33May 9, 2024Updated last year
- RL with Experience Replay☆55Jul 27, 2025Updated 7 months ago
- NeurIPS 2025: Discriminative Constrained Optimization for Reinforcing Large Reasoning Models☆51Feb 3, 2026Updated 3 weeks ago
- ☆36Jul 7, 2025Updated 7 months ago
- PD calibration techniques for LDP portfolios☆10May 29, 2016Updated 9 years ago
- ☆12Oct 29, 2023Updated 2 years ago
- Python powered music controlling webpage with websockets and bottle py (works with spotify, vlc, audacious, and others)☆11Jun 9, 2017Updated 8 years ago
- Optimizing Anytime Reasoning via Budget Relative Policy Optimization