logikon-ai / awesome-deliberative-prompting
Awesome deliberative prompting: How to ask LLMs to produce reliable reasoning and make reason-responsive decisions.
☆114Updated last month
Alternatives and similar repositories for awesome-deliberative-prompting:
Users that are interested in awesome-deliberative-prompting are comparing it to the libraries listed below
- Code repo for "Agent Instructs Large Language Models to be General Zero-Shot Reasoners"☆105Updated 6 months ago
- Github repository for "RAGTruth: A Hallucination Corpus for Developing Trustworthy Retrieval-Augmented Language Models"☆166Updated 3 months ago
- ☆120Updated 9 months ago
- A set of utilities for running few-shot prompting experiments on large-language models☆118Updated last year
- Attribute (or cite) statements generated by LLMs back to in-context information.☆221Updated 5 months ago
- Framework and toolkits for building and evaluating collaborative agents that can work together with humans.☆70Updated last month
- ☆217Updated 7 months ago
- awesome synthetic (text) datasets☆265Updated 5 months ago
- Steer LLM outputs towards a certain topic/subject and enhance response capabilities using activation engineering by adding steering vecto…☆230Updated last month
- Google Deepmind's PromptBreeder for automated prompt engineering implemented in langchain expression language.☆100Updated 7 months ago
- AWM: Agent Workflow Memory☆252Updated 2 months ago
- Official Implementation of "Multi-Head RAG: Solving Multi-Aspect Problems with LLMs"☆202Updated 5 months ago
- [EMNLP 2024 Findings] OneGen: Efficient One-Pass Unified Generation and Retrieval for LLMs.☆147Updated 4 months ago
- augmented LLM with self reflection☆117Updated last year
- ☆143Updated 11 months ago
- 🌍 Repository for "AppWorld: A Controllable World of Apps and People for Benchmarking Interactive Coding Agent", ACL'24 Best Resource Pap…☆170Updated last week
- MiniCheck: Efficient Fact-Checking of LLMs on Grounding Documents [EMNLP 2024]☆139Updated 2 months ago
- Benchmarking LLMs with Challenging Tasks from Real Users☆219Updated 4 months ago
- Code accompanying "How I learned to start worrying about prompt formatting".☆102Updated 5 months ago
- ☆131Updated last year
- An Analytical Evaluation Board of Multi-turn LLM Agents [NeurIPS 2024 Oral]☆297Updated 10 months ago
- A curated list of Human Preference Datasets for LLM fine-tuning, RLHF, and eval.☆350Updated last year
- SelfCheckGPT: Zero-Resource Black-Box Hallucination Detection for Generative Large Language Models☆504Updated 9 months ago
- A simple unified framework for evaluating LLMs☆206Updated 3 weeks ago
- My implementation of "Algorithm of Thoughts: Enhancing Exploration of Ideas in Large Language Models"☆98Updated last year
- This is the official repo for "PromptAgent: Strategic Planning with Language Models Enables Expert-level Prompt Optimization". PromptAgen…☆254Updated 7 months ago
- ☆117Updated 7 months ago
- [EMNLP 2023] Enabling Large Language Models to Generate Text with Citations. Paper: https://arxiv.org/abs/2305.14627☆478Updated 5 months ago
- This repository implements the chain of verification paper by Meta AI☆166Updated last year
- ToolQA, a new dataset to evaluate the capabilities of LLMs in answering challenging questions with external tools. It offers two levels …☆254Updated last year