JacobPfau / fillerTokens
☆60Updated 11 months ago
Alternatives and similar repositories for fillerTokens:
Users that are interested in fillerTokens are comparing it to the libraries listed below
- Code for the arXiv preprint "The Unreasonable Effectiveness of Easy Training Data"☆46Updated last year
- ☆41Updated last month
- ☆163Updated 3 weeks ago
- A Large-Scale, High-Quality Math Dataset for Reinforcement Learning in Language Models☆44Updated last month
- ☆31Updated 2 months ago
- The official repository for SkyLadder: Better and Faster Pretraining via Context Window Scheduling☆28Updated 2 weeks ago
- Long Context Extension and Generalization in LLMs☆50Updated 6 months ago
- ☆96Updated 9 months ago
- ☆74Updated 7 months ago
- Exploration of automated dataset selection approaches at large scales.☆34Updated last month
- NeurIPS 2024 tutorial on LLM Inference☆39Updated 3 months ago
- Replicating O1 inference-time scaling laws☆83Updated 4 months ago
- B-STAR: Monitoring and Balancing Exploration and Exploitation in Self-Taught Reasoners☆75Updated 3 months ago
- [NeurIPS 2024] OlympicArena: Benchmarking Multi-discipline Cognitive Reasoning for Superintelligent AI☆97Updated 3 weeks ago
- This repo is based on https://github.com/jiaweizzhao/GaLore☆26Updated 6 months ago
- Language models scale reliably with over-training and on downstream tasks☆96Updated last year
- Advancing Language Model Reasoning through Reinforcement Learning and Inference Scaling☆95Updated 2 months ago
- Code for Paper: Teaching Language Models to Critique via Reinforcement Learning☆84Updated last month
- Official github repo for the paper "Compression Represents Intelligence Linearly" [COLM 2024]☆130Updated 6 months ago
- Easy-to-Hard Generalization: Scalable Alignment Beyond Human Supervision☆119Updated 6 months ago
- Self-Alignment with Principle-Following Reward Models☆156Updated last year
- Repository for NPHardEval, a quantified-dynamic benchmark of LLMs☆52Updated last year
- Code for PHATGOOSE introduced in "Learning to Route Among Specialized Experts for Zero-Shot Generalization"☆82Updated last year
- [ICLR 2025] LongPO: Long Context Self-Evolution of Large Language Models through Short-to-Long Preference Optimization☆31Updated last month
- Scalable Meta-Evaluation of LLMs as Evaluators☆42Updated last year
- EvaByte: Efficient Byte-level Language Models at Scale☆85Updated 2 weeks ago
- Co-LLM: Learning to Decode Collaboratively with Multiple Language Models☆110Updated 10 months ago