rookie-joe / automatic-lean4-compilation
☆13Updated 3 months ago
Related projects ⓘ
Alternatives and complementary repositories for automatic-lean4-compilation
- ☆20Updated 2 weeks ago
- ☆16Updated 2 weeks ago
- Curation of resources for LLM mathematical reasoning, most of which are screened by @tongyx361 to ensure high quality and accompanied wit…☆84Updated 4 months ago
- Easy-to-Hard Generalization: Scalable Alignment Beyond Human Supervision☆98Updated 2 months ago
- A simple toolkit for benchmarking LLMs on mathematical reasoning tasks. 🧮✨☆105Updated 6 months ago
- ☆51Updated 7 months ago
- [EMNLP 2023] MQuAKE: Assessing Knowledge Editing in Language Models via Multi-Hop Questions☆102Updated 2 months ago
- ☆66Updated 6 months ago
- ☆33Updated 9 months ago
- Code for the paper LEGO-Prover: Neural Theorem Proving with Growing Libraries☆54Updated 8 months ago
- [EMNLP 2024] Source code for the paper "Learning Planning-based Reasoning with Trajectory Collection and Process Rewards Synthesizing".☆43Updated 2 months ago
- Evaluating Mathematical Reasoning Beyond Accuracy☆37Updated 7 months ago
- Collection of papers for scalable automated alignment.☆73Updated last month
- ☆10Updated 4 months ago
- Code for ACL2024 paper - Adversarial Preference Optimization (APO).☆50Updated 5 months ago
- Explore what LLMs are really leanring over SFT☆26Updated 7 months ago
- Implementation of ICML 23 Paper: Specializing Smaller Language Models towards Multi-Step Reasoning.☆125Updated last year
- Synthetic question-answering dataset to formally analyze the chain-of-thought output of large language models on a reasoning task.☆118Updated last month
- ☆24Updated 6 months ago
- Code for the paper <SelfCheck: Using LLMs to Zero-Shot Check Their Own Step-by-Step Reasoning>☆45Updated last year
- ☆25Updated last month
- GenRM-CoT: Data release for verification rationales☆24Updated last month
- [NeurIPS'24] Official code for *🎯DART-Math: Difficulty-Aware Rejection Tuning for Mathematical Problem-Solving*☆79Updated last month
- [ACL'24, Outstanding Paper] Emulated Disalignment: Safety Alignment for Large Language Models May Backfire!☆29Updated 3 months ago
- Official Repo for ICLR 2024 paper MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback by Xingyao Wang*, Ziha…☆104Updated 5 months ago
- BeaverTails is a collection of datasets designed to facilitate research on safety alignment in large language models (LLMs).☆111Updated last year
- ☆25Updated last month
- ☆101Updated 5 months ago
- Non-Autoregressive Math Word Problem Solver with Unified Tree Structure☆11Updated 10 months ago
- ☆31Updated 3 weeks ago