Timothyxxx / Chain-of-ThoughtsPapersLinks
A trend starts from "Chain of Thought Prompting Elicits Reasoning in Large Language Models".
☆2,092Updated 2 years ago
Alternatives and similar repositories for Chain-of-ThoughtsPapers
Users that are interested in Chain-of-ThoughtsPapers are comparing it to the libraries listed below
Sorting:
- [ACL 2023] Reasoning with Language Model Prompting: A Survey☆990Updated 6 months ago
- Paper List for In-context Learning 🌷☆871Updated last year
- Benchmarking large language models' complex reasoning ability with chain-of-thought prompting☆2,762Updated last year
- ☆914Updated last year
- Code for our EMNLP 2023 Paper: "LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models"☆1,217Updated last year
- Official implementation for "Automatic Chain of Thought Prompting in Large Language Models" (stay tuned & more will be updated)☆1,980Updated last year
- The official GitHub page for the survey paper "A Survey on Evaluation of Large Language Models".☆1,584Updated 6 months ago
- Open Academic Research on Improving LLaMA to SOTA LLM☆1,614Updated 2 years ago
- Reading list of hallucination in LLMs. Check out our new survey paper: "Siren’s Song in the AI Ocean: A Survey on Hallucination in Large …☆1,067Updated 2 months ago
- Reading list of Instruction-tuning. A trend starts from Natrural-Instruction (ACL 2022), FLAN (ICLR 2022) and T0 (ICLR 2022).☆768Updated 2 years ago
- ☆2,873Updated 9 months ago
- Aligning Large Language Models with Human: A Survey☆739Updated 2 years ago
- Papers and Datasets on Instruction Tuning and Following. ✨✨✨☆505Updated last year
- ☆923Updated last year
- We unified the interfaces of instruction-tuning data (e.g., CoT data), multiple LLMs and parameter-efficient methods (e.g., lora, p-tunin…☆2,787Updated 2 years ago
- ☆1,330Updated last year
- A plug-and-play library for parameter-efficient-tuning (Delta Tuning)☆1,037Updated last year
- Human preference data for "Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback"☆1,803Updated 5 months ago
- [NIPS2023] RRHF & Wombat☆811Updated 2 years ago
- Secrets of RLHF in Large Language Models Part I: PPO☆1,406Updated last year
- ☆769Updated last year
- Awesome resources for in-context learning and prompt engineering: Mastery of the LLMs such as ChatGPT, GPT-3, and FlanT5, with up-to-date…☆1,656Updated this week
- Must-read papers on prompt-based tuning for pre-trained language models.☆4,296Updated 2 years ago
- An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.☆1,924Updated 4 months ago
- Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback☆1,564Updated 3 weeks ago
- A collection of open-source dataset to train instruction-following LLMs (ChatGPT,LLaMA,Alpaca)☆1,136Updated last year
- 800,000 step-level correctness labels on LLM solutions to MATH problems☆2,076Updated 2 years ago
- Prefix-Tuning: Optimizing Continuous Prompts for Generation☆956Updated last year
- Instruction Tuning with GPT-4☆4,340Updated 2 years ago
- This repository contains a collection of papers and resources on Reasoning in Large Language Models.☆565Updated 2 years ago