avidml / evaluating-LLMs
Creating the tools and data sets necessary to evaluate vulnerabilities in LLMs.
☆23Updated 2 weeks ago
Alternatives and similar repositories for evaluating-LLMs:
Users that are interested in evaluating-LLMs are comparing it to the libraries listed below
- 🤗 Disaggregators: Curated data labelers for in-depth analysis.☆65Updated 2 years ago
- Find and fix bugs in natural language machine learning models using adaptive testing.☆183Updated 10 months ago
- A Python library aimed at dissecting and augmenting NER training data.☆58Updated last year
- Stanford CRFM's initiative to assess potential compliance with the draft EU AI Act☆93Updated last year
- 📚 A curated list of papers & technical articles on AI Quality & Safety☆172Updated last year
- Interpretable and efficient predictors using pre-trained language models. Scikit-learn compatible.☆41Updated 3 weeks ago
- ☆13Updated 2 years ago
- A python package for benchmarking interpretability techniques on Transformers.☆213Updated 6 months ago
- Inquisitive Parrots for Search☆189Updated last year
- Annotated corpus + evaluation metrics for text anonymisation☆55Updated last year
- Command Line Interface for Hugging Face Inference Endpoints☆66Updated 11 months ago
- ☆43Updated last year
- Explainable Zero-Shot Topic Extraction☆62Updated 7 months ago
- [Data + code] ExpertQA : Expert-Curated Questions and Attributed Answers☆126Updated last year
- ☆126Updated 2 years ago
- triple-encoders is a library for contextualizing distributed Sentence Transformers representations.☆14Updated 6 months ago
- Vespa application making an index of the CORD-19 dataset.☆39Updated 2 months ago
- codebase release for EMNLP2023 paper publication☆19Updated last year
- Code associated with the paper "Entropy-based Attention Regularization Frees Unintended Bias Mitigation from Lists"☆48Updated 2 years ago
- ☆65Updated last year
- The Foundation Model Transparency Index☆77Updated 10 months ago
- Tools for managing datasets for governance and training.☆83Updated last month
- A Python package to compute HONEST, a score to measure hurtful sentence completions in language models. Published at NAACL 2021.☆21Updated 2 years ago
- Large-language Model Evaluation framework with Elo Leaderboard and A-B testing☆51Updated 5 months ago
- A framework for few-shot evaluation of autoregressive language models.☆103Updated last year
- Run safety benchmarks against AI models and view detailed reports showing how well they performed.☆83Updated this week
- Using short models to classify long texts☆21Updated 2 years ago
- Notebooks for training universal 0-shot classifiers on many different tasks☆120Updated 3 months ago
- Powerful unsupervised domain adaptation method for dense retrieval. Requires only unlabeled corpus and yields massive improvement: "GPL: …☆330Updated last year
- A library to synthesize text datasets using Large Language Models (LLM)☆151Updated 2 years ago