jxzhangjhu / Awesome-LLM-Prompt-Optimization
Awesome-LLM-Prompt-Optimization: a curated list of advanced prompt optimization and tuning methods in Large Language Models
☆325Updated last year
Alternatives and similar repositories for Awesome-LLM-Prompt-Optimization:
Users that are interested in Awesome-LLM-Prompt-Optimization are comparing it to the libraries listed below
- RewardBench: the first evaluation tool for reward models.☆555Updated last month
- This is a collection of research papers for Self-Correcting Large Language Models with Automated Feedback.☆518Updated 5 months ago
- ☆313Updated 3 weeks ago
- ☆533Updated 3 weeks ago
- A package to evaluate factuality of long-form generation. Original implementation of our EMNLP 2023 paper "FActScore: Fine-grained Atomic…☆340Updated last week
- Official implementation for the paper "DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models"☆481Updated 3 months ago
- Implementation of "RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Horizon Generation".☆230Updated 10 months ago
- LLM hallucination paper list☆314Updated last year
- Code and data for "Lost in the Middle: How Language Models Use Long Contexts"☆340Updated last year
- papers related to LLM-agent that published on top conferences☆314Updated last week
- This repository contains a collection of papers and resources on Reasoning in Large Language Models.☆563Updated last year
- [ACL'24] Selective Reflection-Tuning: Student-Selected Data Recycling for LLM Instruction-Tuning☆354Updated 7 months ago
- SelfCheckGPT: Zero-Resource Black-Box Hallucination Detection for Generative Large Language Models☆509Updated 10 months ago
- This is the official repo for "PromptAgent: Strategic Planning with Language Models Enables Expert-level Prompt Optimization". PromptAgen…☆269Updated 8 months ago
- Github repository for "RAGTruth: A Hallucination Corpus for Developing Trustworthy Retrieval-Augmented Language Models"☆170Updated 4 months ago
- This is the repository of HaluEval, a large-scale hallucination evaluation benchmark for Large Language Models.☆463Updated last year
- Survey on LLM Agents (Published on CoLing 2025)☆228Updated last month
- ICML 2024: Improving Factuality and Reasoning in Language Models through Multiagent Debate☆427Updated last year
- LLMs can generate feedback on their work, use it to improve the output, and repeat this process iteratively.☆679Updated 6 months ago
- Must-read Papers on Knowledge Editing for Large Language Models.☆1,070Updated last month
- List of papers on hallucination detection in LLMs.☆846Updated 2 weeks ago
- [ICML 2024] LESS: Selecting Influential Data for Targeted Instruction Tuning☆435Updated 6 months ago
- Codebase for reproducing the experiments of the semantic uncertainty paper (short-phrase and sentence-length experiments).☆310Updated last year
- LOFT: A 1 Million+ Token Long-Context Benchmark☆190Updated this week
- A curated list of Human Preference Datasets for LLM fine-tuning, RLHF, and eval.☆354Updated last year
- An Analytical Evaluation Board of Multi-turn LLM Agents [NeurIPS 2024 Oral]☆308Updated 11 months ago
- [ICLR 2025] Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing. Your efficient and high-quality synthetic data …☆681Updated last month
- ☆216Updated last year
- The papers are organized according to our survey: Evaluating Large Language Models: A Comprehensive Survey.☆754Updated 11 months ago
- FuseAI Project☆560Updated 3 months ago