aypan17 / machiavelliView external linksLinks
☆146Jul 23, 2025Updated 6 months ago
Alternatives and similar repositories for machiavelli
Users that are interested in machiavelli are comparing it to the libraries listed below
Sorting:
- A library for mechanistic anomaly detection☆22Jan 9, 2025Updated last year
- Pin files for contextual, codebase-level AI assistance.☆16Jul 11, 2024Updated last year
- Repo for the paper on Escalation Risks of AI systems☆44Apr 12, 2024Updated last year
- Experiments with representation engineering☆13Feb 28, 2024Updated last year
- 📚📚📚📚📚📚📚📚📚 Reading everything☆15Sep 12, 2025Updated 5 months ago
- Erasing concepts from neural representations with provable guarantees☆243Jan 27, 2025Updated last year
- WMDP is a LLM proxy benchmark for hazardous knowledge in bio, cyber, and chemical security. We also release code for RMU, an unlearning m…☆158May 29, 2025Updated 8 months ago
- ☆118Jan 19, 2026Updated 3 weeks ago
- ☆16Mar 5, 2024Updated last year
- Our research proposes a novel MoGU framework that improves LLMs' safety while preserving their usability.☆18Jan 14, 2025Updated last year
- TACL 2025: Investigating Adversarial Trigger Transfer in Large Language Models☆19Aug 17, 2025Updated 5 months ago
- The Happy Faces Benchmark☆15Jul 20, 2023Updated 2 years ago
- Aligning AI With Shared Human Values (ICLR 2021)☆315Apr 21, 2023Updated 2 years ago
- LLM experiments done during SERI MATS - focusing on activation steering / interpreting activation spaces☆100Sep 21, 2023Updated 2 years ago
- Code release for "Debating with More Persuasive LLMs Leads to More Truthful Answers"☆124Mar 22, 2024Updated last year
- ☆284Mar 2, 2024Updated last year
- Code for paper "Universal Jailbreak Backdoors from Poisoned Human Feedback"☆66Apr 24, 2024Updated last year
- Contains random samples referenced in the paper "Sleeper Agents: Training Robustly Deceptive LLMs that Persist Through Safety Training".☆127Mar 9, 2024Updated last year
- [USENIX'25] HateBench: Benchmarking Hate Speech Detectors on LLM-Generated Content and Hate Campaigns☆13Mar 1, 2025Updated 11 months ago
- Representation Engineering: A Top-Down Approach to AI Transparency☆947Aug 14, 2024Updated last year
- In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination Mitigation (ICML 2024)☆62Mar 30, 2024Updated last year
- A collection of different ways to implement accessing and modifying internal model activations for LLMs☆21Oct 18, 2024Updated last year
- Formalizing stochastic doubly-efficient debate☆118Oct 8, 2024Updated last year
- ☆11Mar 13, 2023Updated 2 years ago
- Instant Neural Graphics Primitives from scratch, zero dependencies. Learning by doing.☆10Aug 18, 2023Updated 2 years ago
- Inference-Time Intervention: Eliciting Truthful Answers from a Language Model☆570Jan 28, 2025Updated last year
- [ICLR 2025] General-purpose activation steering library☆142Sep 18, 2025Updated 4 months ago
- Investigating the generalization behavior of LM probes trained to predict truth labels: (1) from one annotator to another, and (2) from e…☆28May 23, 2024Updated last year
- A curated list of resources on Reinforcement Learning with Verifiable Rewards (RLVR) and the reasoning capability boundary of Large Langu…☆85Dec 12, 2025Updated 2 months ago
- Keeping language models honest by directly eliciting knowledge encoded in their activations.☆217Jan 26, 2026Updated 2 weeks ago
- ☆12Jul 8, 2023Updated 2 years ago
- Codebase for Mechanistic Mode Connectivity☆13Jul 14, 2023Updated 2 years ago
- code of paper "Defending Against Alignment-Breaking Attacks via Robustly Aligned LLM"☆14Nov 17, 2023Updated 2 years ago
- ☆11Oct 16, 2023Updated 2 years ago
- 中文原生等级化代码能力测试基准☆15Apr 11, 2024Updated last year
- ☆33Jan 25, 2026Updated 3 weeks ago
- Repository for NPHardEval, a quantified-dynamic benchmark of LLMs☆63Mar 26, 2024Updated last year
- ☆267Oct 1, 2024Updated last year
- Search, Verify and Feedback: Towards Next Generation Post-training Paradigm of Foundation Models via Verifier Engineering☆63Dec 5, 2024Updated last year