NextWordDev / psychoevalsLinks
Repository for PsychoEvals - a framework for LLM security, psychoanalysis, and moderation.
☆18Updated 2 years ago
Alternatives and similar repositories for psychoevals
Users that are interested in psychoevals are comparing it to the libraries listed below
Sorting:
- PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to a…☆449Updated last year
- Large Language Models Meet NL2Code: A Survey☆35Updated last year
- 📚 A curated list of papers & technical articles on AI Quality & Safety☆199Updated 9 months ago
- Analyzing and scoring reasoning traces of LLMs☆47Updated last year
- repo for the paper titled “CodeGen4Libs: A Two-Stage Approach for Library-Oriented Code Generation”☆14Updated 2 years ago
- ☆54Updated 10 months ago
- Official repo for GPTFUZZER : Red Teaming Large Language Models with Auto-Generated Jailbreak Prompts☆563Updated last year
- Code for the AAAI 2023 paper "CodeAttack: Code-based Adversarial Attacks for Pre-Trained Programming Language Models☆34Updated 2 years ago
- [NeurIPS 2024] Evaluation harness for SWT-Bench, a benchmark for evaluating LLM repository-level test-generation☆67Updated last month
- Enhancing AI Software Engineering with Repository-level Code Graph☆246Updated 9 months ago
- Data and code for "DocPrompting: Generating Code by Retrieving the Docs" @ICLR 2023☆251Updated 2 years ago
- Official repo for Customized but Compromised: Assessing Prompt Injection Risks in User-Designed GPTs☆30Updated 2 years ago
- The official implementation of our NAACL 2024 paper "A Wolf in Sheep’s Clothing: Generalized Nested Jailbreak Prompts can Fool Large Lang…☆151Updated 4 months ago
- [NeurIPS 2023 D&B] Code repository for InterCode benchmark https://arxiv.org/abs/2306.14898☆235Updated last year
- Awesome deliberative prompting: How to ask LLMs to produce reliable reasoning and make reason-responsive decisions.☆120Updated 11 months ago
- Can It Edit? Evaluating the Ability of Large Language Models to Follow Code Editing Instructions☆48Updated 4 months ago
- Official code for the paper "CodeChain: Towards Modular Code Generation Through Chain of Self-revisions with Representative Sub-modules"☆48Updated 2 months ago
- Whispers in the Machine: Confidentiality in Agentic Systems☆41Updated last month
- ☆68Updated last year
- Code and Data for: Reading Between the Lines: Modeling User Behavior and Costs in AI-Assisted Programming☆33Updated last year
- ☆36Updated 2 years ago
- Curation of prompts that are known to be adversarial to large language models☆188Updated 2 years ago
- A set of utilities for running few-shot prompting experiments on large-language models☆126Updated 2 years ago
- autoredteam: code for training models that automatically red team other language models☆15Updated 2 years ago
- ☆102Updated last year
- We introduce new zero-shot prompting magic words that improves the reasoning ability of language models: panel discussion!☆170Updated last year
- APIBench is a benchmark for evaluating the performance of API recommendation approaches released in the paper "Revisiting, Benchmarking a…☆65Updated 2 years ago
- LLM Platform Security: Applying a Systematic Evaluation Framework to OpenAI's ChatGPT Plugins☆29Updated last year
- Code and data of the EMNLP 2022 paper "Why Should Adversarial Perturbations be Imperceptible? Rethink the Research Paradigm in Adversaria…☆69Updated 2 years ago
- This repository contains the replication package of our paper "Assessing the Security of GitHub Copilot’s Generated Code - A Targeted Rep…☆10Updated 2 years ago