lechmazur / deceptionLinks
Benchmark evaluating LLMs on their ability to create and resist disinformation. Includes comprehensive testing across major models (Claude, GPT-4, Gemini, Llama, etc.) with standardized evaluation metrics.
☆31Updated 10 months ago
Alternatives and similar repositories for deception
Users that are interested in deception are comparing it to the libraries listed below
Sorting:
- Multi-Agent Step Race Benchmark: Assessing LLM Collaboration and Deception Under Pressure. A multi-player “step-race” that challenges LLM…☆85Updated last month
- LLM Divergent Thinking Creativity Benchmark. LLMs generate 25 unique words that start with a given letter with no connections to each oth…☆35Updated 10 months ago
- Thematic Generalization Benchmark: measures how effectively various LLMs can infer a narrow or specific "theme" (category/rule) from a sm…☆63Updated 4 months ago
- Benchmark that evaluates LLMs using 759 NYT Connections puzzles extended with extra trick words☆190Updated last month
- A subset of jailbreaks automatically discovered by the Haize Labs haizing suite.☆100Updated 9 months ago
- Glyphs, acting as collaboratively defined symbols linking related concepts, add a layer of multidimensional semantic richness to user-AI …☆56Updated 11 months ago
- ☆38Updated 7 months ago
- A preprint version of our recent research on the capability of frontier AI systems to do self-replication☆59Updated last year
- Locally hosted AI Agent Python Tool To Generate Novel Research Hypothesis + Titles + Abstracts☆27Updated 8 months ago
- Adding a multi-text multi-speaker script (diffe) that is based on a script from asiff00 on issue 61 for Sesame: A Conversational Speech G…☆26Updated 9 months ago
- An easy-to-understand framework for LLM samplers that rewind and revise generated tokens☆150Updated 2 weeks ago
- ☆62Updated 6 months ago
- A Python library to orchestrate LLMs in a neural network-inspired structure☆52Updated last year
- Neuroengine is a service to share LLMs in the form of a webchat and API.☆45Updated last year
- ☆15Updated last year
- never forget anything again! combine AI and intelligent tooling for a local knowledge base to track catalogue, annotate, and plan for you…☆37Updated last year
- Hallucinations (Confabulations) Document-Based Benchmark for RAG. Includes human-verified questions and answers.☆241Updated 5 months ago
- Digital Red Queen: Adversarial Program Evolution in Core War with LLMs☆158Updated 2 weeks ago
- OpenPipe Reinforcement Learning Experiments☆32Updated 10 months ago
- ☆30Updated last year
- A simple experiment on letting two local LLM have a conversation about anything!☆112Updated last year
- klmbr - a prompt pre-processing technique to break through the barrier of entropy while generating text with LLMs☆86Updated last year
- Lego for GRPO☆30Updated 8 months ago
- Experimental sampler to make LLMs more creative☆31Updated 2 years ago
- Modified Beam Search with periodical restart☆12Updated last year
- ☆24Updated last year
- Outputs from the Deep Writer☆16Updated last year
- ☆50Updated last year
- ☆88Updated 2 months ago
- Groq-powered MAD: The first work to explore Multi-Agent Debate with Large Language Models :D☆12Updated last year