☆229Feb 23, 2021Updated 5 years ago
Alternatives and similar repositories for real-toxicity-prompts
Users that are interested in real-toxicity-prompts are comparing it to the libraries listed below
Sorting:
- This repository contains the data and code introduced in the paper "CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Maske…☆130Mar 1, 2024Updated 2 years ago
- Dataset associated with "BOLD: Dataset and Metrics for Measuring Biases in Open-Ended Language Generation" paper☆87Mar 2, 2021Updated 5 years ago
- This repository contains the code for "Self-Diagnosis and Self-Debiasing: A Proposal for Reducing Corpus-Based Bias in NLP".☆89Aug 20, 2021Updated 4 years ago
- [NeurIPS 2024 D&B] Evaluating Copyright Takedown Methods for Language Models☆17Jul 17, 2024Updated last year
- Röttger et al. (NAACL 2024): "XSTest: A Test Suite for Identifying Exaggerated Safety Behaviours in Large Language Models"☆129Feb 24, 2025Updated last year
- code associated with ACL 2021 DExperts paper☆118May 24, 2023Updated 2 years ago
- TruthfulQA: Measuring How Models Imitate Human Falsehoods☆890Jan 16, 2025Updated last year
- Aligning AI With Shared Human Values (ICLR 2021)☆316Apr 21, 2023Updated 2 years ago
- This repo contains the code for generating the ToxiGen dataset, published at ACL 2022.☆345Jun 17, 2024Updated last year
- Official Implementation of "Learning to Refuse: Towards Mitigating Privacy Risks in LLMs"☆10Dec 13, 2024Updated last year
- Repository for the Bias Benchmark for QA dataset.☆139Jan 8, 2024Updated 2 years ago
- This is the official code for the paper "Vaccine: Perturbation-aware Alignment for Large Language Models" (NeurIPS2024)☆49Jan 15, 2026Updated last month
- ICLR2024 Paper. Showing properties of safety tuning and exaggerated safety.☆93May 9, 2024Updated last year
- ☆164Sep 2, 2024Updated last year
- Data for evaluating gender bias in coreference resolution systems.☆81May 14, 2019Updated 6 years ago
- Human preference data for "Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback"☆1,824Jun 17, 2025Updated 8 months ago
- Do-Not-Answer: A Dataset for Evaluating Safeguards in LLMs☆319Jun 7, 2024Updated last year
- Butler 是一个用于自动化服务管理和任务调度的工具项目。☆16Mar 2, 2026Updated last week
- Efficient and Effective Weight-Ensembling Mixture of Experts for Multi-Task Model Merging. Arxiv, 2024.☆16Oct 28, 2024Updated last year
- GeDi: Generative Discriminator Guided Sequence Generation☆211Jun 16, 2025Updated 8 months ago
- ☆19Jun 21, 2025Updated 8 months ago
- A library for mechanistic anomaly detection☆22Jan 9, 2025Updated last year
- ☆17Dec 21, 2023Updated 2 years ago
- Official repo for NeurIPS'24 paper "WAGLE: Strategic Weight Attribution for Effective and Modular Unlearning in Large Language Models"☆19Dec 16, 2024Updated last year
- Using GPT-3 to detect hate speech that contains sexist and racist content☆24Nov 11, 2025Updated 4 months ago
- Function Vectors in Large Language Models (ICLR 2024)☆192Apr 17, 2025Updated 10 months ago
- WMDP is a LLM proxy benchmark for hazardous knowledge in bio, cyber, and chemical security. We also release code for RMU, an unlearning m…☆160May 29, 2025Updated 9 months ago
- A dataset of alignment research and code to reproduce it☆78Jun 22, 2023Updated 2 years ago
- Data set for LREC 2020 paper "I Feel Offended, Don't Be Abusive!"☆18Sep 23, 2023Updated 2 years ago
- ☆60Aug 22, 2024Updated last year
- StereoSet: Measuring stereotypical bias in pretrained language models☆199Dec 8, 2022Updated 3 years ago
- ☆32Aug 9, 2024Updated last year
- A Comprehensive Assessment of Trustworthiness in GPT Models☆314Sep 16, 2024Updated last year
- ☆44Oct 1, 2024Updated last year
- A Python package to compute HONEST, a score to measure hurtful sentence completions in language models. Published at NAACL 2021.☆21Apr 8, 2025Updated 11 months ago
- ☆26Sep 5, 2024Updated last year
- 🌏 UI component library for the future, based on WebComponent.☆23Nov 12, 2024Updated last year
- ☆147Jul 23, 2025Updated 7 months ago
- ☆30Aug 2, 2024Updated last year