fyzhang1 / OblivionisLinks
[AAAI 2026] Official Repository for "Oblivionis: A Lightweight Learning and Unlearning Framework for Federated Large Language Models"
☆34Updated last month
Alternatives and similar repositories for Oblivionis
Users that are interested in Oblivionis are comparing it to the libraries listed below
Sorting:
- To Think or Not to Think: Exploring the Unthinking Vulnerability in Large Reasoning Models☆32Updated 7 months ago
- Repo for SemStamp (NAACL2024) and k-SemStamp (ACL2024)☆27Updated last year
- [USENIX Security 2025] PoisonedRAG: Knowledge Corruption Attacks to Retrieval-Augmented Generation of Large Language Models☆222Updated last month
- [USENIX Security 2025] SOFT: Selective Data Obfuscation for Protecting LLM Fine-tuning against Membership Inference Attacks☆16Updated 3 months ago
- Some Conferences' accepted paper lists (including Network, Security, AI) 顶会近五年论文列表☆49Updated 2 months ago
- Code for paper "SrcMarker: Dual-Channel Source Code Watermarking via Scalable Code Transformations" (IEEE S&P 2024)☆33Updated last year
- [CCS 2024] Optimization-based Prompt Injection Attack to LLM-as-a-Judge☆37Updated 3 months ago
- Code and data for paper "Can LLM Watermarks Robustly Prevent Unauthorized Knowledge Distillation?". (ACL 2025 Main)☆20Updated 6 months ago
- ☆13Updated 5 months ago
- Code for paper "The Philosopher’s Stone: Trojaning Plugins of Large Language Models"☆25Updated last year
- The code for paper "The Good and The Bad: Exploring Privacy Issues in Retrieval-Augmented Generation (RAG)", exploring the privacy risk o…☆63Updated 11 months ago
- Safety at Scale: A Comprehensive Survey of Large Model Safety☆216Updated last month
- ☆21Updated last year
- Code and data for paper "A Semantic Invariant Robust Watermark for Large Language Models" accepted by ICLR 2024.☆37Updated last year
- [ICLR 2025] Dissecting adversarial robustness of multimodal language model agents☆122Updated 10 months ago
- multi-bit language model watermarking (NAACL 24)☆17Updated last year
- ☆21Updated 2 months ago
- ☆22Updated last year
- Source code of paper "An Unforgeable Publicly Verifiable Watermark for Large Language Models" accepted by ICLR 2024☆34Updated last year
- ☆15Updated this week
- SG-Bench: Evaluating LLM Safety Generalization Across Diverse Tasks and Prompt Types☆23Updated last year
- A list of recent papers about adversarial learning☆277Updated this week
- This is the code repository for "Uncovering Safety Risks of Large Language Models through Concept Activation Vector"☆47Updated 2 months ago
- ☆26Updated last year
- A Survey on Jailbreak Attacks and Defenses against Multimodal Generative Models☆297Updated last month
- Agent Security Bench (ASB)☆167Updated 2 months ago
- [NeurIPS 2025] BackdoorLLM: A Comprehensive Benchmark for Backdoor Attacks and Defenses on Large Language Models☆262Updated 2 months ago
- ☆53Updated 7 months ago
- ☆118Updated 11 months ago
- [ICLR24] Official Repo of BadChain: Backdoor Chain-of-Thought Prompting for Large Language Models☆44Updated last year