pillowsofwind / Course-Correction
The official GitHub repo for the paper "Course-Correction: Safety Alignment Using Synthetic Preferences"
☆19Updated last month
Related projects: ⓘ
- ☆27Updated 3 months ago
- Code & Data for our Paper "Alleviating Hallucinations of Large Language Models through Induced Hallucinations"☆56Updated 6 months ago
- The official GitHub repo for the survey paper "Knowledge Conflicts for LLMs: A Survey"☆62Updated 3 weeks ago
- Official repository for ICML 2024 paper "On Prompt-Driven Safeguarding for Large Language Models"☆64Updated 2 weeks ago
- R-Judge: Benchmarking Safety Risk Awareness for LLM Agents☆57Updated last month
- [ICLR'24] RAIN: Your Language Models Can Align Themselves without Finetuning☆79Updated 3 months ago
- To Forget or Not? Towards Practical Knowledge Unlearning for Large Language Models☆16Updated last month
- ☆61Updated 3 months ago
- Multilingual safety benchmark for Large Language Models☆21Updated 2 weeks ago
- ☆32Updated 10 months ago
- [EMNLP 2023] MQuAKE: Assessing Knowledge Editing in Language Models via Multi-Hop Questions☆96Updated last week
- Knowledge Circuits in Pretrained Transformers☆46Updated this week
- Github repository for "FELM: Benchmarking Factuality Evaluation of Large Language Models"☆54Updated 8 months ago
- 【ACL 2024】 SALAD benchmark & MD-Judge☆81Updated this week
- RWKU: Benchmarking Real-World Knowledge Unlearning for Large Language Models☆50Updated 2 months ago
- Weak-to-Strong Jailbreaking on Large Language Models☆62Updated 6 months ago
- ☆46Updated 2 weeks ago
- Semi-Parametric Editing with a Retrieval-Augmented Counterfactual Model☆59Updated last year
- Official code implementation of SKU, Accepted by ACL 2024 Findings☆11Updated 4 months ago
- Official repository for paper "Weak-to-Strong Extrapolation Expedites Alignment"☆62Updated 3 months ago
- The Paper List on Data Contamination for Large Language Models Evaluation.☆46Updated this week
- A Mechanistic Understanding of Alignment Algorithms: A Case Study on DPO and Toxicity.☆46Updated last month
- Shadow Alignment: The Ease of Subverting Safely-Aligned Language Models☆22Updated 11 months ago
- ☆13Updated 6 months ago
- The repository of the project "Fine-tuning Large Language Models with Sequential Instructions", code base comes from open-instruct and LA…☆30Updated 2 months ago
- Evaluating the Ripple Effects of Knowledge Editing in Language Models☆45Updated 5 months ago
- Restore safety in fine-tuned language models through task arithmetic☆25Updated 5 months ago
- Mostly recording papers about models' trustworthy applications. Intending to include topics like model evaluation & analysis, security, c…☆18Updated last year
- Official Code for Paper: Assessing the Brittleness of Safety Alignment via Pruning and Low-Rank Modifications☆55Updated 2 months ago
- ☆42Updated 5 months ago