haizelabs / dspy-redteamLinks
Red-Teaming Language Models with DSPy
☆203Updated 5 months ago
Alternatives and similar repositories for dspy-redteam
Users that are interested in dspy-redteam are comparing it to the libraries listed below
Sorting:
- A subset of jailbreaks automatically discovered by the Haize Labs haizing suite.☆95Updated 3 months ago
- Inference-time scaling for LLMs-as-a-judge.☆267Updated 3 weeks ago
- ☆24Updated 9 months ago
- Sphynx Hallucination Induction☆53Updated 6 months ago
- ⚖️ Awesome LLM Judges ⚖️☆108Updated 3 months ago
- ☆127Updated last month
- Guardrails for secure and robust agent development☆327Updated last week
- Collection of evals for Inspect AI☆198Updated this week
- A DSPy-based implementation of the tree of thoughts method (Yao et al., 2023) for generating persuasive arguments☆87Updated 10 months ago
- ☆102Updated this week
- Vivaria is METR's tool for running evaluations and conducting agent elicitation research.☆103Updated last week
- Code for the paper "Fishing for Magikarp"☆162Updated 2 months ago
- Attribute (or cite) statements generated by LLMs back to in-context information.☆268Updated 9 months ago
- A framework for optimizing DSPy programs with RL☆96Updated this week
- A repository of Language Model Vulnerabilities and Exposures (LVEs).☆113Updated last year
- A better way of testing, inspecting, and analyzing AI Agent traces.☆39Updated 3 weeks ago
- ☆130Updated 4 months ago
- Just a bunch of benchmark logs for different LLMs☆119Updated last year
- Functional Benchmarks and the Reasoning Gap☆88Updated 10 months ago
- Contains random samples referenced in the paper "Sleeper Agents: Training Robustly Deceptive LLMs that Persist Through Safety Training".☆111Updated last year
- The fastest Trust Layer for AI Agents☆140Updated 2 months ago
- ☆71Updated 9 months ago
- Approximation of the Claude 3 tokenizer by inspecting generation stream☆131Updated last year
- ☆34Updated last month
- TapeAgents is a framework that facilitates all stages of the LLM Agent development lifecycle☆289Updated last week
- Curation of prompts that are known to be adversarial to large language models☆184Updated 2 years ago
- Open source interpretability artefacts for R1.☆157Updated 3 months ago
- ☆78Updated 9 months ago
- ☆182Updated 4 months ago
- Initiative to evaluate and rank the most popular LLMs across common task types based on their propensity to hallucinate.☆113Updated last week