alipay / private_llmLinks
☆35Updated last year
Alternatives and similar repositories for private_llm
Users that are interested in private_llm are comparing it to the libraries listed below
Sorting:
- Shepherd: A foundational framework enabling federated instruction tuning for large language models☆247Updated 2 years ago
- [ICLR'24 Spotlight] DP-OPT: Make Large Language Model Your Privacy-Preserving Prompt Engineer☆46Updated last year
- The official implement of paper "Does Federated Learning Really Need Backpropagation?"☆23Updated 2 years ago
- Hide and Seek (HaS): A Framework for Prompt Privacy Protection☆48Updated 2 years ago
- Federated Learning for LLMs.☆235Updated last week
- Implementation for PrE-Text: Training Language Models on Private Federated Data in the Age of LLMs☆24Updated last year
- Official repo for the paper: Recovering Private Text in Federated Learning of Language Models (in NeurIPS 2022)☆61Updated 2 years ago
- R-Judge: Benchmarking Safety Risk Awareness for LLM Agents (EMNLP Findings 2024)☆91Updated 5 months ago
- FedJudge: Federated Legal Large Language Model☆36Updated last year
- A MoE impl for PyTorch, [ATC'23] SmartMoE☆71Updated 2 years ago
- ShieldLM: Empowering LLMs as Aligned, Customizable and Explainable Safety Detectors [EMNLP 2024 Findings]☆214Updated last year
- [ICLR'24] RAIN: Your Language Models Can Align Themselves without Finetuning☆99Updated last year
- Codebase for decoding compressed trust.☆24Updated last year
- BeaverTails is a collection of datasets designed to facilitate research on safety alignment in large language models (LLMs).☆163Updated 2 years ago
- A survey of privacy problems in Large Language Models (LLMs). Contains summary of the corresponding paper along with relevant code☆68Updated last year
- This repository provides an original implementation of Detecting Pretraining Data from Large Language Models by *Weijia Shi, *Anirudh Aji…☆232Updated 2 years ago
- Official implementation of Privacy Implications of Retrieval-Based Language Models (EMNLP 2023). https://arxiv.org/abs/2305.14888☆36Updated last year
- ☆12Updated 2 years ago
- [EMNLP 2023] Lion: Adversarial Distillation of Proprietary Large Language Models☆212Updated last year
- Offsite-Tuning: Transfer Learning without Full Model☆380Updated last year
- ☆160Updated 9 months ago
- Privacy-Preserving Prompt Tuning for Large Language Model☆28Updated last year
- A simple GPT-based evaluation tool for multi-aspect, interpretable assessment of LLMs.☆87Updated last year
- LLM Unlearning☆177Updated 2 years ago
- Official github repo for SafetyBench, a comprehensive benchmark to evaluate LLMs' safety. [ACL 2024]☆260Updated 3 months ago
- S-Eval: Towards Automated and Comprehensive Safety Evaluation for Large Language Models☆99Updated 2 weeks ago
- Shadow Alignment: The Ease of Subverting Safely-Aligned Language Models☆32Updated 2 years ago
- On Memorization of Large Language Models in Logical Reasoning☆72Updated 7 months ago
- Counting-Stars (★)☆83Updated 4 months ago
- ☆47Updated 7 months ago