koalazf99 / nanoverl
Collections of RLxLM experiments using minimal codes
☆12Updated 2 months ago
Alternatives and similar repositories for nanoverl:
Users that are interested in nanoverl are comparing it to the libraries listed below
- ☆13Updated 9 months ago
- BeHonest: Benchmarking Honesty in Large Language Models☆31Updated 8 months ago
- [AAAI 2025 oral] Evaluating Mathematical Reasoning Beyond Accuracy☆60Updated 4 months ago
- GSM-Plus: Data, Code, and Evaluation for Enhancing Robust Mathematical Reasoning in Math Word Problems.☆59Updated 9 months ago
- ☆59Updated 7 months ago
- The rule-based evaluation subset and code implementation of Omni-MATH☆19Updated 3 months ago
- [EMNLP 2022] TaCube: Pre-computing Data Cubes for Answering Numerical-Reasoning Questions over Tabular Data☆17Updated last year
- The repository of the project "Fine-tuning Large Language Models with Sequential Instructions", code base comes from open-instruct and LA…☆29Updated 4 months ago
- Safety-J: Evaluating Safety with Critique☆16Updated 8 months ago
- Analyzing LLM Alignment via Token distribution shift☆16Updated last year
- LongProc: Benchmarking Long-Context Language Models on Long Procedural Generation☆23Updated 2 weeks ago
- ☆16Updated 5 months ago
- ☆29Updated 3 months ago
- ☆12Updated 5 months ago
- M-STAR (Multimodal Self-Evolving TrAining for Reasoning) Project. Diving into Self-Evolving Training for Multimodal Reasoning☆56Updated 3 months ago
- ☆44Updated 5 months ago
- Evaluation utilities based on SymPy.☆16Updated 4 months ago
- Code and data used in the paper: "Training on Incorrect Synthetic Data via RL Scales LLM Math Reasoning Eight-Fold"☆30Updated 10 months ago
- The code of “Improving Weak-to-Strong Generalization with Scalable Oversight and Ensemble Learning”☆16Updated last year
- Evaluate the Quality of Critique☆34Updated 10 months ago
- [ICML'24] TroVE: Inducing Verifiable and Efficient Toolboxes for Solving Programmatic Tasks☆26Updated 7 months ago
- [ACL 2024] Code for "MoPS: Modular Story Premise Synthesis for Open-Ended Automatic Story Generation"☆35Updated 9 months ago
- [ACL 2024 Findings] CriticBench: Benchmarking LLMs for Critique-Correct Reasoning☆24Updated last year
- Code and data for "Living in the Moment: Can Large Language Models Grasp Co-Temporal Reasoning?" (ACL 2024)☆32Updated 9 months ago
- ☆73Updated 10 months ago
- The official repository of the Omni-MATH benchmark.☆80Updated 3 months ago
- ☆65Updated last year
- The official code repository for PRMBench.☆72Updated 2 months ago
- Extending context length of visual language models☆11Updated 4 months ago
- ☆12Updated last year