Reading list of Instruction-tuning. A trend starts from Natrural-Instruction (ACL 2022), FLAN (ICLR 2022) and T0 (ICLR 2022).
β766Jul 20, 2023Updated 2 years ago
Alternatives and similar repositories for Instruction-Tuning-Papers
Users that are interested in Instruction-Tuning-Papers are comparing it to the libraries listed below
Sorting:
- Paper List for In-context Learning π·β875Oct 8, 2024Updated last year
- A trend starts from "Chain of Thought Prompting Elicits Reasoning in Large Language Models".β2,100Oct 5, 2023Updated 2 years ago
- Papers and Datasets on Instruction Tuning and Following. β¨β¨β¨β508Apr 4, 2024Updated last year
- Aligning pretrained language models with instruction data generated by themselves.β4,576Mar 27, 2023Updated 2 years ago
- Benchmarking large language models' complex reasoning ability with chain-of-thought promptingβ2,766Aug 4, 2024Updated last year
- Expanding natural instructionsβ1,035Dec 11, 2023Updated 2 years ago
- Paper collections of methods that using language to interact with environment, including interact with real world, simulated world or WWWβ¦β129Jul 26, 2023Updated 2 years ago
- A collection of open-source dataset to train instruction-following LLMs (ChatGPT,LLaMA,Alpaca)β1,143Jan 4, 2024Updated 2 years ago
- Must-read papers on prompt-based tuning for pre-trained language models.β4,295Jul 17, 2023Updated 2 years ago
- This repository contains a collection of papers and resources on Reasoning in Large Language Models.β567Nov 13, 2023Updated 2 years ago
- [ACL 2023] Reasoning with Language Model Prompting: A Surveyβ996May 21, 2025Updated 9 months ago
- A curated list of reinforcement learning with human feedback resources (continually updated)β4,301Dec 9, 2025Updated 2 months ago
- A modular RL library to fine-tune language models to human preferencesβ2,378Mar 1, 2024Updated last year
- ICML'2022: Black-Box Tuning for Language-Model-as-a-Service & EMNLP'2022: BBTv2: Towards a Gradient-Free Future with Large Language Modelβ¦β271Nov 8, 2022Updated 3 years ago
- Paper List for Contrastive Learning for Natural Language Processingβ571Apr 27, 2023Updated 2 years ago
- β15Aug 18, 2022Updated 3 years ago
- Paper collection on building and evaluating language model agents via executable language groundingβ365Apr 29, 2024Updated last year
- β917Jul 24, 2024Updated last year
- Toolkit for creating, sharing and using natural language prompts.β2,996Oct 23, 2023Updated 2 years ago
- β1,560Feb 20, 2026Updated last week
- Must-read papers, related blogs and API tools on the pre-training and tuning methods for ChatGPT.β326Aug 10, 2023Updated 2 years ago
- Instruction Tuning with GPT-4β4,342Jun 11, 2023Updated 2 years ago
- Datasets for Instruction Tuning of Large Language Modelsβ261Nov 30, 2023Updated 2 years ago
- Resource, Evaluation and Detection Papers for ChatGPTβ456Mar 21, 2024Updated last year
- A curated list of awesome instruction tuning datasets, models, papers and repositories.β348Jun 12, 2023Updated 2 years ago
- Human preference data for "Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback"β1,816Jun 17, 2025Updated 8 months ago
- [ICML 2023] Exploring the Benefits of Training Expert Language Models over Instruction Tuningβ98Apr 26, 2023Updated 2 years ago
- From Chain-of-Thought prompting to OpenAI o1 and DeepSeek-R1 πβ3,550May 7, 2025Updated 9 months ago
- [NIPS2023] RRHF & Wombatβ809Sep 22, 2023Updated 2 years ago
- A framework for few-shot evaluation of language models.β11,478Feb 15, 2026Updated last week
- Materials for ACL-2022 tutorial: Knowledge-Augmented Methods for Natural Language Processingβ286Aug 8, 2022Updated 3 years ago
- Aligning Large Language Models with Human: A Surveyβ741Sep 11, 2023Updated 2 years ago
- π©Ί A collection of ChatGPT evaluation reports on various bechmarks.β50Mar 28, 2023Updated 2 years ago
- π€ PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.β20,678Updated this week
- A framework for human-readable prompt-based method with large language models. Specially designed for researchers. (Deprecated, check outβ¦β131Feb 25, 2023Updated 3 years ago
- This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.β552Mar 10, 2024Updated last year
- [ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parametersβ5,936Mar 14, 2024Updated last year
- This is the oficial repository for "Parameter-Efficient Multi-task Tuning via Attentional Mixtures of Soft Prompts" (EMNLP 2022)β104Dec 1, 2022Updated 3 years ago
- Open Academic Research on Improving LLaMA to SOTA LLMβ1,611Aug 30, 2023Updated 2 years ago