ZetangForward / CMD-Context-aware-Model-self-Detoxification
CMD: a framework for Context-aware Model self-Detoxification (EMNLP2024 Main)
☆14Updated last month
Related projects ⓘ
Alternatives and complementary repositories for CMD-Context-aware-Model-self-Detoxification
- Code & Data for our Paper "Alleviating Hallucinations of Large Language Models through Induced Hallucinations"☆60Updated 8 months ago
- A Survey on the Honesty of Large Language Models☆46Updated last month
- BeHonest: Benchmarking Honesty in Large Language Models☆30Updated 3 months ago
- The official repository of "Improving Large Language Models via Fine-grained Reinforcement Learning with Minimum Editing Constraint"☆33Updated 10 months ago
- Paper list and datasets for the paper: A Survey on Data Selection for LLM Instruction Tuning☆33Updated 9 months ago
- L-CITEEVAL: DO LONG-CONTEXT MODELS TRULY LEVERAGE CONTEXT FOR RESPONDING?☆19Updated last month
- Evaluating the Ripple Effects of Knowledge Editing in Language Models☆50Updated 7 months ago
- ☆65Updated 6 months ago
- Code for paper 'Are We Falling in a Middle-Intelligence Trap? An Analysis and Mitigation of the Reversal Curse'☆12Updated 3 months ago
- ☆33Updated last year
- Semi-Parametric Editing with a Retrieval-Augmented Counterfactual Model☆65Updated 2 years ago
- Multilingual safety benchmark for Large Language Models☆24Updated 2 months ago
- One Network, Many Masks: Towards More Parameter-Efficient Transfer Learning☆38Updated last year
- Official code for ICML 2024 paper on Persona In-Context Learning (PICLe)☆21Updated 4 months ago
- [NeurIPS 2023] Github repository for "Composing Parameter-Efficient Modules with Arithmetic Operations"☆58Updated 11 months ago
- RWKU: Benchmarking Real-World Knowledge Unlearning for Large Language Models. NeurIPS 2024☆62Updated last month
- ☆36Updated 10 months ago
- ☆27Updated last year
- ☆71Updated 10 months ago
- We introduce ScaleQuest, a scalable, novel and cost-effective data synthesis method to unleash the reasoning capability of LLMs.☆51Updated 3 weeks ago
- [NeurIPS 2024 Oral] Aligner: Efficient Alignment by Learning to Correct☆120Updated last week
- [NeurIPS 2024] Knowledge Circuits in Pretrained Transformers☆75Updated last month
- [EMNLP 2023] MQuAKE: Assessing Knowledge Editing in Language Models via Multi-Hop Questions☆102Updated 2 months ago
- Github repository for "FELM: Benchmarking Factuality Evaluation of Large Language Models" (NeurIPS 2023)☆54Updated 10 months ago
- The code of “Improving Weak-to-Strong Generalization with Scalable Oversight and Ensemble Learning”☆15Updated 8 months ago
- Repo for the EMNLP'24 Paper "Dual-Space Knowledge Distillation for Large Language Models".☆37Updated 2 weeks ago
- A Survey of Hallucination in Large Foundation Models☆50Updated 10 months ago
- [ICLR'24] RAIN: Your Language Models Can Align Themselves without Finetuning☆84Updated 5 months ago
- Official repository for paper "Weak-to-Strong Extrapolation Expedites Alignment"☆68Updated 5 months ago
- OpenBA-V2: 3B LLM (Large Language Model) with T5 architecture, utilizing model pruning technique and continuing pretraining from OpenBA-1…☆23Updated 6 months ago