chuanyang-Zheng / Progressive-Hint
This is the official implementation of "Progressive-Hint Prompting Improves Reasoning in Large Language Models"
☆205Updated last year
Alternatives and similar repositories for Progressive-Hint:
Users that are interested in Progressive-Hint are comparing it to the libraries listed below
- FireAct: Toward Language Agent Fine-tuning☆265Updated last year
- Codes and Data for Scaling Relationship on Learning Mathematical Reasoning with Large Language Models☆245Updated 5 months ago
- Official implementation of paper "Cumulative Reasoning With Large Language Models" (https://arxiv.org/abs/2308.04371)☆288Updated 5 months ago
- Multi-agent Social Simulation + Efficient, Effective, and Stable alternative of RLHF. Code for the paper "Training Socially Aligned Langu…☆348Updated last year
- Generative Judge for Evaluating Alignment☆228Updated last year
- a Fine-tuned LLaMA that is Good at Arithmetic Tasks☆177Updated last year
- Code and data for "MAmmoTH: Building Math Generalist Models through Hybrid Instruction Tuning" (ICLR 2024)☆360Updated 5 months ago
- ☆137Updated last year
- 🐋 An unofficial implementation of Self-Alignment with Instruction Backtranslation.☆136Updated 7 months ago
- ☆139Updated 7 months ago
- [EMNLP 2024] LongAlign: A Recipe for Long Context Alignment of LLMs☆241Updated 2 months ago
- ToolkenGPT: Augmenting Frozen Language Models with Massive Tools via Tool Embeddings - NeurIPS 2023 (oral)☆256Updated 10 months ago
- Data and Code for Program of Thoughts (TMLR 2023)☆259Updated 9 months ago
- ACL 2024 | LooGLE: Long Context Evaluation for Long-Context Language Models☆175Updated 4 months ago
- An opensource ChatBot built with ExpertPrompting which achieves 96% of ChatGPT's capability.☆298Updated last year
- Official codebase for "SelFee: Iterative Self-Revising LLM Empowered by Self-Feedback Generation"☆226Updated last year
- Offical Repo for "Programming Every Example: Lifting Pre-training Data Quality Like Experts at Scale"☆217Updated this week
- GPT-Fathom is an open-source and reproducible LLM evaluation suite, benchmarking 10+ leading open-source and closed-source LLMs as well a…☆349Updated 10 months ago
- Large Language Models Are Reasoning Teachers (ACL 2023)☆320Updated last year
- ☆172Updated last year
- Unofficial implementation of AlpaGasus☆90Updated last year
- Paper collection on building and evaluating language model agents via executable language grounding☆347Updated 9 months ago
- ToolBench, an evaluation suite for LLM tool manipulation capabilities.☆150Updated 11 months ago
- [ACL'24 Outstanding] Data and code for L-Eval, a comprehensive long context language models evaluation benchmark☆370Updated 7 months ago
- All available datasets for Instruction Tuning of Large Language Models☆242Updated last year
- Challenging BIG-Bench Tasks and Whether Chain-of-Thought Can Solve Them☆459Updated 7 months ago
- Evaluating LLMs' multi-round chatting capability via assessing conversations generated by two LLM instances.☆144Updated last year
- [ACL 2024] AutoAct: Automatic Agent Learning from Scratch for QA via Self-Planning☆207Updated last month
- ☆304Updated 5 months ago
- Self-Alignment with Principle-Following Reward Models☆154Updated 11 months ago