lakeraai / canica
A text embedding viewer for the Jupyter environment
☆19Updated last year
Alternatives and similar repositories for canica:
Users that are interested in canica are comparing it to the libraries listed below
- A benchmark for prompt injection detection systems.☆96Updated 2 weeks ago
- A repository of Language Model Vulnerabilities and Exposures (LVEs).☆108Updated 11 months ago
- [Corca / ML] Automatically solved Gandalf AI with LLM☆48Updated last year
- ☆43Updated 2 years ago
- Risks and targets for assessing LLMs & LLM vulnerabilities☆30Updated 8 months ago
- ☆40Updated 6 months ago
- Finding trojans in aligned LLMs. Official repository for the competition hosted at SaTML 2024.☆109Updated 8 months ago
- A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents.☆96Updated this week
- PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to a…☆333Updated 11 months ago
- Turning Gandalf against itself. Use LLMs to automate playing Lakera Gandalf challenge without needing to set up an account with a platfor…☆29Updated last year
- Dropbox LLM Security research code and results☆220Updated 9 months ago
- Project LLM Verification Standard☆38Updated 10 months ago
- ATLAS tactics, techniques, and case studies data☆56Updated 4 months ago
- A framework-less approach to robust agent development.☆154Updated this week
- Implementation of BEAST adversarial attack for language models (ICML 2024)☆79Updated 9 months ago
- Code to break Llama Guard☆31Updated last year
- AI Verify☆137Updated this week
- The repository contains the code for analysing the leakage of personally identifiable (PII) information from the output of next word pred…☆88Updated 6 months ago
- LLM security and privacy☆47Updated 4 months ago
- ⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs☆350Updated last year
- ☆34Updated 3 months ago
- ☆13Updated 8 months ago
- Payloads for Attacking Large Language Models☆74Updated 7 months ago
- Dataset for the Tensor Trust project☆36Updated 11 months ago
- Tree of Attacks (TAP) Jailbreaking Implementation☆99Updated last year
- This repository provides implementation to formalize and benchmark Prompt Injection attacks and defenses☆172Updated last month
- Code used to run the platform for the LLM CTF colocated with SaTML 2024☆26Updated 11 months ago
- Run safety benchmarks against AI models and view detailed reports showing how well they performed.☆79Updated this week
- TaskTracker is an approach to detecting task drift in Large Language Models (LLMs) by analysing their internal activations. It provides a…☆43Updated 2 months ago
- Adversarial Attacks on GPT-4 via Simple Random Search [Dec 2023]☆43Updated 9 months ago