GenAISHAP / TokenSHAPLinks
TokenSHAP: Explain individual token importance in large language model prompts with SHAP values. Gain insights, debug models, detect biases, and enhance transparency effortlessly
β60Updated this week
Alternatives and similar repositories for TokenSHAP
Users that are interested in TokenSHAP are comparing it to the libraries listed below
Sorting:
- π¦ Unitxt is a Python library for enterprise-grade evaluation of AI performance, offering the world's largest catalog of tools and data β¦β212Updated last week
- A mechanistic approach for understanding and detecting factual errors of large language models.β49Updated last year
- Toolkit for attaching, training, saving and loading of new heads for transformer modelsβ292Updated 9 months ago
- Generalist and Lightweight Model for Text Classificationβ166Updated last week
- PyTorch library for Active Fine-Tuningβ95Updated 2 months ago
- β111Updated 10 months ago
- Notebooks for training universal 0-shot classifiers on many different tasksβ137Updated 11 months ago
- Truth Forest: Toward Multi-Scale Truthfulness in Large Language Models through Intervention without Tuningβ46Updated last year
- Pre-train Static Word Embeddingsβ93Updated 3 months ago
- Erasing concepts from neural representations with provable guaranteesβ239Updated 10 months ago
- β56Updated 10 months ago
- β49Updated last year
- β87Updated last week
- A framework for standardizing evaluations of large foundation models, beyond single-score reporting and rankings.β173Updated 2 weeks ago
- β120Updated last year
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absoluteβ¦β51Updated last year
- β251Updated 3 months ago
- β144Updated 3 months ago
- The Foundation Model Transparency Indexβ83Updated this week
- Efficiently find the best-suited language model (LM) for your NLP taskβ132Updated 4 months ago
- β159Updated last year
- Attribute (or cite) statements generated by LLMs back to in-context information.β305Updated last year
- β124Updated last year
- β69Updated last year
- β148Updated last year
- β79Updated last year
- EvalAssist is an open-source project that simplifies using large language models as evaluators (LLM-as-a-Judge) of the output of other laβ¦β92Updated 2 weeks ago
- [NeurIPS 2024] Goldfish Loss: Mitigating Memorization in Generative LLMsβ93Updated last year
- Mixing Language Models with Self-Verification and Meta-Verificationβ110Updated last year
- Evaluation of neuro-symbolic enginesβ40Updated last year