shizhediao / active-promptLinks
Source code for the paper "Active Prompting with Chain-of-Thought for Large Language Models"
☆243Updated last year
Alternatives and similar repositories for active-prompt
Users that are interested in active-prompt are comparing it to the libraries listed below
Sorting:
- ToolQA, a new dataset to evaluate the capabilities of LLMs in answering challenging questions with external tools. It offers two levels …☆272Updated last year
- ☆284Updated last year
- Code and data accompanying our paper on arXiv "Faithful Chain-of-Thought Reasoning".☆161Updated last year
- Data and Code for Program of Thoughts [TMLR 2023]☆279Updated last year
- [NeurIPS 2023] This is the code for the paper `Large Language Model as Attributed Training Data Generator: A Tale of Diversity and Bias`.☆150Updated last year
- [EMNLP 2023] The CoT Collection: Improving Zero-shot and Few-shot Learning of Language Models via Chain-of-Thought Fine-Tuning☆244Updated last year
- ☆172Updated 2 years ago
- [EMNLP 2023] Enabling Large Language Models to Generate Text with Citations. Paper: https://arxiv.org/abs/2305.14627☆490Updated 9 months ago
- A Survey of Attributions for Large Language Models☆205Updated 10 months ago
- [NeurIPS 2023] Codebase for the paper: "Guiding Large Language Models with Directional Stimulus Prompting"☆111Updated 2 years ago
- Generative Judge for Evaluating Alignment☆244Updated last year
- paper list on reasoning in NLP☆190Updated 3 months ago
- Data and code for FreshLLMs (https://arxiv.org/abs/2310.03214)☆364Updated last week
- This is the repo for the paper Shepherd -- A Critic for Language Model Generation☆219Updated last year
- [ICLR 2023] Code for the paper "Binding Language Models in Symbolic Languages"☆319Updated last year
- Accompanying repo for the RLPrompt paper☆334Updated last year
- FireAct: Toward Language Agent Fine-tuning☆280Updated last year
- Implementation of the paper: "Answering Questions by Meta-Reasoning over Multiple Chains of Thought"☆96Updated last year
- This is the repository of HaluEval, a large-scale hallucination evaluation benchmark for Large Language Models.☆488Updated last year
- ToolBench, an evaluation suite for LLM tool manipulation capabilities.☆154Updated last year
- Code for Arxiv 2023: Improving Language Model Negociation with Self-Play and In-Context Learning from AI Feedback☆207Updated 2 years ago
- ToolkenGPT: Augmenting Frozen Language Models with Massive Tools via Tool Embeddings - NeurIPS 2023 (oral)☆262Updated last year
- Source Code of Paper "GPTScore: Evaluate as You Desire"☆252Updated 2 years ago
- Prod Env☆423Updated last year
- Repository for Interleaving Retrieval with Chain-of-Thought Reasoning for Knowledge-Intensive Multi-Step Questions, ACL23☆222Updated last year
- All available datasets for Instruction Tuning of Large Language Models☆254Updated last year
- [Preprint] Learning to Filter Context for Retrieval-Augmented Generaton☆192Updated last year
- Code and data for "Lost in the Middle: How Language Models Use Long Contexts"☆351Updated last year
- ☆183Updated 5 months ago
- ☆139Updated last year