lxx0628 / Prompting-Framework-SurveyLinks
A curated list of awesome publications and researchers on prompting framework updated and maintained by The Intelligent System Security (IS2).
☆84Updated 9 months ago
Alternatives and similar repositories for Prompting-Framework-Survey
Users that are interested in Prompting-Framework-Survey are comparing it to the libraries listed below
Sorting:
- Open Implementations of LLM Analyses☆107Updated last year
- Mixing Language Models with Self-Verification and Meta-Verification☆109Updated 10 months ago
- Official repo of Respond-and-Respond: data, code, and evaluation☆104Updated last year
- Astraios: Parameter-Efficient Instruction Tuning Code Language Models☆62Updated last year
- ☆43Updated last year
- ☆80Updated 7 months ago
- Codebase accompanying the Summary of a Haystack paper.☆79Updated last year
- CodeSage: Code Representation Learning At Scale (ICLR 2024)☆113Updated last year
- a curated list of the role of small models in the LLM era☆105Updated last year
- ☆48Updated last year
- Code repo for "Agent Instructs Large Language Models to be General Zero-Shot Reasoners"☆116Updated last week
- A list of LLM benchmark frameworks.☆71Updated last year
- 🔧 Compare how Agent systems perform on several benchmarks. 📊🚀☆102Updated 2 months ago
- [ICLR'24 Spotlight] A language model (LM)-based emulation framework for identifying the risks of LM agents with tool use☆167Updated last year
- Data preparation code for CrystalCoder 7B LLM☆45Updated last year
- ☆101Updated last year
- Model, Code & Data for the EMNLP'23 paper "Making Large Language Models Better Data Creators"☆134Updated 2 years ago
- AuditNLG: Auditing Generative AI Language Modeling for Trustworthiness☆101Updated 9 months ago
- A set of utilities for running few-shot prompting experiments on large-language models☆125Updated 2 years ago
- EMNLP 2024 "Re-reading improves reasoning in large language models". Simply repeating the question to get bidirectional understanding for…☆27Updated 10 months ago
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute…☆50Updated last year
- LLMs as Collaboratively Edited Knowledge Bases☆45Updated last year
- Data and evaluation scripts for "CodePlan: Repository-level Coding using LLMs and Planning", FSE 2024☆75Updated last year
- Meta-CoT: Generalizable Chain-of-Thought Prompting in Mixed-task Scenarios with Large Language Models☆99Updated 2 years ago
- LangCode - Improving alignment and reasoning of large language models (LLMs) with natural language embedded program (NLEP).☆46Updated 2 years ago
- ☆79Updated last year
- Formal-LLM: Integrating Formal Language and Natural Language for Controllable LLM-based Agents☆128Updated last year
- [NeurIPS 2024] Evaluation harness for SWT-Bench, a benchmark for evaluating LLM repository-level test-generation☆58Updated last month
- RAGElo is a set of tools that helps you selecting the best RAG-based LLM agents by using an Elo ranker☆120Updated this week
- Open-source Self-Instruction Tuning Code LLM☆169Updated 2 years ago