jxzhangjhu / Awesome-LLM-Prompt-Optimization
Awesome-LLM-Prompt-Optimization: a curated list of advanced prompt optimization and tuning methods in Large Language Models
☆311Updated last year
Alternatives and similar repositories for Awesome-LLM-Prompt-Optimization:
Users that are interested in Awesome-LLM-Prompt-Optimization are comparing it to the libraries listed below
- RewardBench: the first evaluation tool for reward models.☆536Updated last month
- This is a collection of research papers for Self-Correcting Large Language Models with Automated Feedback.☆513Updated 5 months ago
- This is the official repo for "PromptAgent: Strategic Planning with Language Models Enables Expert-level Prompt Optimization". PromptAgen…☆254Updated 7 months ago
- Official implementation for the paper "DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models"☆476Updated 2 months ago
- Evaluate your LLM's response with Prometheus and GPT4 💯☆893Updated 2 weeks ago
- AWM: Agent Workflow Memory☆252Updated 2 months ago
- A curated collection of LLM reasoning and planning resources, including key papers, limitations, benchmarks, and additional learning mate…☆249Updated last month
- SelfCheckGPT: Zero-Resource Black-Box Hallucination Detection for Generative Large Language Models☆504Updated 9 months ago
- An Analytical Evaluation Board of Multi-turn LLM Agents [NeurIPS 2024 Oral]☆297Updated 10 months ago
- Building a comprehensive and handy list of papers for GUI agents☆269Updated 2 weeks ago
- Generative Representational Instruction Tuning☆613Updated 2 weeks ago
- Implementation of "RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Horizon Generation".☆230Updated 10 months ago
- Codebase for reproducing the experiments of the semantic uncertainty paper (short-phrase and sentence-length experiments).☆295Updated 11 months ago
- [ACL'24] Selective Reflection-Tuning: Student-Selected Data Recycling for LLM Instruction-Tuning☆350Updated 6 months ago
- ICML 2024: Improving Factuality and Reasoning in Language Models through Multiagent Debate☆417Updated last year
- This is the repository of HaluEval, a large-scale hallucination evaluation benchmark for Large Language Models.☆453Updated last year
- ☆277Updated 3 weeks ago
- LLM hallucination paper list☆312Updated last year
- [NeurIPS 2024 Spotlight] Buffer of Thoughts: Thought-Augmented Reasoning with Large Language Models☆614Updated last week
- ReSearch: Learning to Reason with Search for LLMs via Reinforcement Learning☆395Updated this week
- Generative Judge for Evaluating Alignment☆232Updated last year
- [ICLR 2025] Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing. Your efficient and high-quality synthetic data …☆664Updated 2 weeks ago
- Automatic evals for LLMs☆346Updated this week
- ☆284Updated 9 months ago
- Model Merging in LLMs, MLLMs, and Beyond: Methods, Theories, Applications and Opportunities. arXiv:2408.07666.☆347Updated this week
- [ICML 2024] LESS: Selecting Influential Data for Targeted Instruction Tuning☆424Updated 5 months ago
- A simple unified framework for evaluating LLMs☆206Updated 3 weeks ago
- The code and data for "MMLU-Pro: A More Robust and Challenging Multi-Task Language Understanding Benchmark" [NeurIPS 2024]☆212Updated last month
- Github repository for "RAGTruth: A Hallucination Corpus for Developing Trustworthy Retrieval-Augmented Language Models"☆166Updated 3 months ago
- Representation Engineering: A Top-Down Approach to AI Transparency☆811Updated 7 months ago