AI4LIFE-GROUP / LLM_ExplainerLinks
Code for paper: Are Large Language Models Post Hoc Explainers?
☆33Updated last year
Alternatives and similar repositories for LLM_Explainer
Users that are interested in LLM_Explainer are comparing it to the libraries listed below
Sorting:
- [EMNLP 2023] Poisoning Retrieval Corpora by Injecting Adversarial Passages https://arxiv.org/abs/2310.19156☆37Updated last year
- Conformal Language Modeling☆32Updated last year
- OpenDataVal: a Unified Benchmark for Data Valuation in Python (NeurIPS 2023)☆99Updated 7 months ago
- ☆36Updated last year
- A repository for summaries of recent explainable AI/Interpretable ML approaches☆83Updated 11 months ago
- Using Explanations as a Tool for Advanced LLMs☆67Updated last year
- DataInf: Efficiently Estimating Data Influence in LoRA-tuned LLMs and Diffusion Models (ICLR 2024)☆76Updated 11 months ago
- ☆100Updated last year
- ☆148Updated 2 years ago
- Influence Analysis and Estimation - Survey, Papers, and Taxonomy☆82Updated last year
- A Python Data Valuation Package☆30Updated 2 years ago
- Uncertainty quantification for in-context learning of large language models☆16Updated last year
- The dataset and code for the ICLR 2024 paper "Can LLM-Generated Misinformation Be Detected?"☆76Updated 10 months ago
- A reproduced PyTorch implementation of the Adversarially Reweighted Learning (ARL) model, originally presented in "Fairness without Demog…☆20Updated 4 years ago
- ☆32Updated last year
- Code for Language-Interfaced FineTuning for Non-Language Machine Learning Tasks.☆130Updated 10 months ago
- A Mechanistic Understanding of Alignment Algorithms: A Case Study on DPO and Toxicity.☆78Updated 6 months ago
- code repo for ICLR 2024 paper "Can LLMs Express Their Uncertainty? An Empirical Evaluation of Confidence Elicitation in LLMs"☆132Updated last year
- ☆55Updated 2 years ago
- ☆45Updated 7 months ago
- A simple PyTorch implementation of influence functions.☆91Updated last year
- Fairness in LLMs resources☆32Updated last month
- Data and code for the Corr2Cause paper (ICLR 2024)☆111Updated last year
- A resource repository for representation engineering in large language models☆136Updated 10 months ago
- ☆20Updated last year
- ☆174Updated last year
- FFB: A Fair Fairness Benchmark for In-Processing Group Fairness Methods.☆30Updated last year
- The TABLET benchmark for evaluating instruction learning with LLMs for tabular prediction.☆21Updated 2 years ago
- 🤫 Code and benchmark for our ICLR 2024 spotlight paper: "Can LLMs Keep a Secret? Testing Privacy Implications of Language Models via Con…☆44Updated last year
- ☆13Updated 2 years ago