embeddings-benchmark / mtebLinks
MTEB: Massive Text Embedding Benchmark
☆3,106Updated this week
Alternatives and similar repositories for mteb
Users that are interested in mteb are comparing it to the libraries listed below
Sorting:
- This includes the original implementation of SELF-RAG: Learning to Retrieve, Generate and Critique through self-reflection by Akari Asai,…☆2,315Updated last year
- Distilabel is a framework for synthetic data and AI feedback for engineers who need fast, reliable and scalable pipelines based on verifi…☆3,084Updated 2 weeks ago
- A Heterogeneous Benchmark for Information Retrieval. Easy to use, evaluate your models across 15+ diverse IR datasets.☆2,064Updated 3 months ago
- Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.☆2,877Updated this week
- Fast lexical search implementing BM25 in Python using Numpy, Numba and Scipy☆1,477Updated this week
- [ACL 2023] One Embedder, Any Task: Instruction-Finetuned Text Embeddings☆2,022Updated last year
- ColBERT: state-of-the-art neural search (SIGIR'20, TACL'21, NeurIPS'21, NAACL'22, CIKM'22, ACL'23, EMNLP'23)☆3,770Updated 3 months ago
- Enforce the output format (JSON Schema, Regex etc) of a language model☆1,986Updated 5 months ago
- Doing simple retrieval from LLM models at various context lengths to measure accuracy☆2,167Updated last year
- Efficient Retrieval Augmentation and Generation Framework☆1,766Updated 3 weeks ago
- ☆2,121Updated last year
- Bringing BERT into modernity via both architecture changes and scaling☆1,627Updated 7 months ago
- The official implementation of RAPTOR: Recursive Abstractive Processing for Tree-Organized Retrieval☆1,572Updated last year
- A blazing fast inference solution for text embeddings models☆4,476Updated this week
- An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.☆1,940Updated 6 months ago
- Code for 'LLM2Vec: Large Language Models Are Secretly Powerful Text Encoders'☆1,648Updated 2 months ago
- Lighteval is your all-in-one toolkit for evaluating LLMs across multiple backends☆2,291Updated 2 weeks ago
- Holistic Evaluation of Language Models (HELM) is an open source Python framework created by the Center for Research on Foundation Models …☆2,662Updated this week
- A lightweight, low-dependency, unified API to use all common reranking and cross-encoder models.☆1,594Updated last month
- Measuring Massive Multitask Language Understanding | ICLR 2021☆1,550Updated 2 years ago
- Toolkit for creating, sharing and using natural language prompts.☆2,997Updated 2 years ago
- Aligning pretrained language models with instruction data generated by themselves.☆4,573Updated 2 years ago
- Evaluate your LLM's response with Prometheus and GPT4 💯☆1,043Updated 9 months ago
- Pyserini is a Python toolkit for reproducible information retrieval research with sparse and dense representations.☆2,013Updated last week
- SPLADE: sparse neural search (SIGIR21, SIGIR22)☆978Updated last year
- Retrieval and Retrieval-augmented LLMs☆11,256Updated last month
- ⚡FlashRAG: A Python Toolkit for Efficient RAG Research (WWW2025 Resource)☆3,309Updated 2 months ago
- A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)☆3,151Updated 2 months ago
- Automated Evaluation of RAG Systems☆687Updated 10 months ago
- Data and tools for generating and inspecting OLMo pre-training data.☆1,403Updated 3 months ago