ronigold / TokenSHAPLinks
TokenSHAP: Explain individual token importance in large language model prompts with SHAP values. Gain insights, debug models, detect biases, and enhance transparency effortlessly
☆45Updated 2 months ago
Alternatives and similar repositories for TokenSHAP
Users that are interested in TokenSHAP are comparing it to the libraries listed below
Sorting:
- Top papers related to LLM-based agent evaluation☆68Updated 2 weeks ago
- A mechanistic approach for understanding and detecting factual errors of large language models.☆46Updated 11 months ago
- A repository containing the code for translating popular LLM benchmarks to German.☆25Updated last year
- ☆58Updated 3 weeks ago
- Tensor Extraction of Latent Features (T-ELF). Within T-ELF's arsenal are non-negative matrix and tensor factorization solutions, equipped…☆19Updated 3 weeks ago
- The Official Repository for "Bring Your Own Data! Self-Supervised Evaluation for Large Language Models"☆108Updated last year
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute…☆49Updated 10 months ago
- A small library of LLM judges☆205Updated 2 weeks ago
- Mixing Language Models with Self-Verification and Meta-Verification☆104Updated 5 months ago
- Truth Forest: Toward Multi-Scale Truthfulness in Large Language Models through Intervention without Tuning☆46Updated last year
- PyTorch library for Active Fine-Tuning☆79Updated 3 months ago
- The first dense retrieval model that can be prompted like an LM☆73Updated 3 weeks ago
- ☆23Updated last month
- ☆69Updated last year
- Stanford NLP Python library for benchmarking the utility of LLM interpretability methods☆92Updated this week
- Code for reproducing our paper "Not All Language Model Features Are Linear"☆75Updated 6 months ago
- Code for In-context Vectors: Making In Context Learning More Effective and Controllable Through Latent Space Steering☆175Updated 3 months ago
- The Granite Guardian models are designed to detect risks in prompts and responses.☆85Updated 2 months ago
- An introduction to LLM Sampling☆78Updated 5 months ago
- Functional Benchmarks and the Reasoning Gap☆86Updated 8 months ago
- Combining Base and Instruction-Tuned Language Models for Better Synthetic Data Generation☆31Updated 3 months ago
- Erasing concepts from neural representations with provable guarantees☆228Updated 4 months ago
- ☆45Updated 4 months ago
- Public code repo for paper "SaySelf: Teaching LLMs to Express Confidence with Self-Reflective Rationales"☆106Updated 8 months ago
- Code for "On Measuring Faithfulness of Natural Language Explanations"☆19Updated 10 months ago
- ☆27Updated last month
- Run safety benchmarks against AI models and view detailed reports showing how well they performed.☆92Updated this week
- Pre-train Static Word Embeddings☆76Updated this week
- Using open source LLMs to build synthetic datasets for direct preference optimization☆63Updated last year
- Code for the EMNLP 2024 paper "Detecting and Mitigating Contextual Hallucinations in Large Language Models Using Only Attention Maps"☆126Updated 9 months ago