junxu-ai / LLM_fairnessLinks
Collection of papers, tools, datasets for fairness of LLM
☆13Updated 9 months ago
Alternatives and similar repositories for LLM_fairness
Users that are interested in LLM_fairness are comparing it to the libraries listed below
Sorting:
- A resource repository for machine unlearning in large language models☆435Updated last month
- The lastest paper about detection of LLM-generated text and code☆275Updated last month
- ☆140Updated last year
- A survey on harmful fine-tuning attack for large language model☆193Updated 2 weeks ago
- ☆18Updated last year
- ☆20Updated last year
- awesome papers in LLM interpretability☆522Updated last month
- A reading list for large models safety, security, and privacy (including Awesome LLM Security, Safety, etc.).☆1,553Updated last week
- A curated list of safety-related papers, articles, and resources focused on Large Language Models (LLMs). This repository aims to provide…☆1,490Updated last week
- Must-read Papers on Knowledge Editing for Large Language Models.☆1,111Updated last week
- ☆25Updated last month
- ☆50Updated last year
- [NeurIPS 2024 D&B] DetectRL: Benchmarking LLM-Generated Text Detection in Real-World Scenarios☆27Updated 7 months ago
- ☆12Updated 6 months ago
- [USENIX Security 2025] PoisonedRAG: Knowledge Corruption Attacks to Retrieval-Augmented Generation of Large Language Models☆172Updated 4 months ago
- Papers and resources related to the security and privacy of LLMs 🤖☆517Updated last month
- Awesome papers involving LLMs in Social Science.☆515Updated this week
- Official code for the paper Towards Fully Exploiting LLM Internal States to Enhance Knowledge Boundary Perception. The code is based on t…☆15Updated 3 weeks ago
- ☆582Updated last month
- UP-TO-DATE LLM Watermark paper. 🔥🔥🔥☆343Updated 7 months ago
- [ICML 2024] TrustLLM: Trustworthiness in Large Language Models☆581Updated 3 weeks ago
- ☆144Updated 10 months ago
- Continuously updated list of related resources for generative LLMs like GPT and their analysis and detection.☆222Updated last month
- Reading list of hallucination in LLMs. Check out our new survey paper: "Siren’s Song in the AI Ocean: A Survey on Hallucination in Large …☆1,026Updated 8 months ago
- ☆15Updated last year
- [ICLR 2024] The official implementation of our ICLR2024 paper "AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language M…☆360Updated 5 months ago
- Paper list of misinformation research using (multi-modal) large language models, i.e., (M)LLMs.☆269Updated 7 months ago
- [NAACL2024] Attacks, Defenses and Evaluations for LLM Conversation Safety: A Survey☆105Updated 11 months ago
- Toolkit for evaluating the trustworthiness of generative foundation models.☆106Updated 3 weeks ago
- Repo for SemStamp (NAACL2024) and k-SemStamp (ACL2024)☆21Updated 7 months ago