baixuechunzi / llm-implicit-bias
☆14Updated 2 months ago
Alternatives and similar repositories for llm-implicit-bias
Users that are interested in llm-implicit-bias are comparing it to the libraries listed below
Sorting:
- [ICML 2025] Weak-to-Strong Jailbreaking on Large Language Models☆74Updated 2 weeks ago
- This repository contains the code and data for the paper "SelfIE: Self-Interpretation of Large Language Model Embeddings" by Haozhe Chen,…☆48Updated 5 months ago
- Code for the paper <SelfCheck: Using LLMs to Zero-Shot Check Their Own Step-by-Step Reasoning>☆49Updated last year
- Source code and data for ADEPT: A DEbiasing PrompT Framework (AAAI-23).☆14Updated 5 months ago
- LLM experiments done during SERI MATS - focusing on activation steering / interpreting activation spaces☆92Updated last year
- Recent papers on (1) Psychology of LLMs; (2) Biases in LLMs.☆48Updated last year
- The official repo of paper "Self-Control of LLM Behaviors by Compressing Suffix Gradient into Prefix Controller"☆18Updated 9 months ago
- code for EMNLP 2024 paper: Neuron-Level Knowledge Attribution in Large Language Models☆32Updated 6 months ago
- Code release for "Debating with More Persuasive LLMs Leads to More Truthful Answers"☆105Updated last year
- The Paper List on Data Contamination for Large Language Models Evaluation.☆93Updated last month
- Paper list for the survey "Combating Misinformation in the Age of LLMs: Opportunities and Challenges" and the initiative "LLMs Meet Misin…☆103Updated 6 months ago
- ☆170Updated last year
- Repo for the research paper "SecAlign: Defending Against Prompt Injection with Preference Optimization"☆45Updated last month
- Official Repository for ACL 2024 Paper SafeDecoding: Defending against Jailbreak Attacks via Safety-Aware Decoding☆130Updated 9 months ago
- [ICLR 2025] Cheating Automatic LLM Benchmarks: Null Models Achieve High Win Rates (Oral)☆78Updated 6 months ago
- [NeurIPS 2024] How do Large Language Models Handle Multilingualism?☆34Updated 6 months ago
- Codes and datasets for the paper Measuring and Enhancing Trustworthiness of LLMs in RAG through Grounded Attributions and Learning to Ref…☆56Updated 2 months ago
- [ACL'24] Chain of Thought (CoT) is significant in improving the reasoning abilities of large language models (LLMs). However, the correla…☆46Updated last week
- ☆69Updated 3 months ago
- Official Repository for The Paper: Safety Alignment Should Be Made More Than Just a Few Tokens Deep☆119Updated 3 weeks ago
- [ICLR'24] RAIN: Your Language Models Can Align Themselves without Finetuning☆93Updated 11 months ago
- Röttger et al. (NAACL 2024): "XSTest: A Test Suite for Identifying Exaggerated Safety Behaviours in Large Language Models"☆97Updated 2 months ago
- ☆58Updated 10 months ago
- A resource repository for representation engineering in large language models☆120Updated 6 months ago
- 【ACL 2024】 SALAD benchmark & MD-Judge☆145Updated 2 months ago
- A Mechanistic Understanding of Alignment Algorithms: A Case Study on DPO and Toxicity.☆72Updated 2 months ago
- ☆46Updated last week
- Code associated with Tuning Language Models by Proxy (Liu et al., 2024)☆109Updated last year
- ☆43Updated last year
- awesome SAE papers☆27Updated 2 months ago