invariantlabs-ai / invariantLinks
Guardrails for secure and robust agent development
☆334Updated last month
Alternatives and similar repositories for invariant
Users that are interested in invariant are comparing it to the libraries listed below
Sorting:
- A better way of testing, inspecting, and analyzing AI Agent traces.☆40Updated last month
- Red-Teaming Language Models with DSPy☆208Updated 6 months ago
- ☆142Updated 2 months ago
- The fastest Trust Layer for AI Agents☆142Updated 2 months ago
- Sandboxed code execution for AI agents, locally or on the cloud. Massively parallel, easy to extend. Powering SWE-agent and more.☆286Updated last week
- Inference-time scaling for LLMs-as-a-judge.☆276Updated last month
- DeepTeam is a framework to red team LLMs and LLM systems.☆636Updated last week
- A repository of Language Model Vulnerabilities and Exposures (LVEs).☆113Updated last year
- TaskTracker is an approach to detecting task drift in Large Language Models (LLMs) by analysing their internal activations. It provides a…☆64Updated 2 weeks ago
- A security scanner for your LLM agentic workflows☆681Updated 3 weeks ago
- Collection of evals for Inspect AI☆211Updated this week
- Enhancing AI Software Engineering with Repository-level Code Graph☆205Updated 4 months ago
- A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents.☆244Updated 2 weeks ago
- ☆576Updated 2 weeks ago
- Python SDK for running evaluations on LLM generated responses☆292Updated 2 months ago
- CodeSage: Code Representation Learning At Scale (ICLR 2024)☆112Updated 10 months ago
- LLM proxy to observe and debug what your AI agents are doing.☆46Updated last month
- ☆117Updated 2 weeks ago
- Sphynx Hallucination Induction☆53Updated 6 months ago
- An agent benchmark with tasks in a simulated software company.☆534Updated this week
- PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to a…☆406Updated last year
- ☆71Updated 10 months ago
- ⚖️ Awesome LLM Judges ⚖️☆122Updated 3 months ago
- An open-source compliance-centered evaluation framework for Generative AI models☆161Updated this week
- A Text-Based Environment for Interactive Debugging☆256Updated this week
- A plugin-based gateway that orchestrates other MCPs and allows developers to build upon it enterprise-grade agents.☆265Updated last month
- Lightweight and portable LLM sandbox runtime (code interpreter) Python library.☆451Updated this week
- Vivaria is METR's tool for running evaluations and conducting agent elicitation research.☆108Updated last week
- Kura is a simple reproduction of the CLIO paper which uses language models to label user behaviour before clustering them based on embedd…☆296Updated last month
- 🤖 Headless IDE for AI agents☆200Updated 4 months ago