huggingface / evaluateLinks
π€ Evaluate: A library for easily evaluating machine learning models and datasets.
β2,247Updated last week
Alternatives and similar repositories for evaluate
Users that are interested in evaluate are comparing it to the libraries listed below
Sorting:
- Efficient few-shot learning with Sentence Transformersβ2,512Updated 2 months ago
- A Unified Library for Parameter-Efficient and Modular Transfer Learningβ2,726Updated last month
- π Accelerate inference and training of π€ Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimizationβ¦β2,953Updated this week
- Holistic Evaluation of Language Models (HELM) is an open source Python framework created by the Center for Research on Foundation Models β¦β2,310Updated this week
- β1,527Updated 2 weeks ago
- Model explainability that works seamlessly with π€ transformers. Explain your transformers model in just 2 lines of code.β1,355Updated last year
- A modular RL library to fine-tune language models to human preferencesβ2,320Updated last year
- A Heterogeneous Benchmark for Information Retrieval. Easy to use, evaluate your models across 15+ diverse IR datasets.β1,850Updated 3 weeks ago
- β2,835Updated 3 weeks ago
- PyTorch extensions for high performance and large scale training.β3,336Updated 2 months ago
- Accessible large language models via k-bit quantization for PyTorch.β7,167Updated this week
- Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.β2,434Updated last week
- Toolkit for creating, sharing and using natural language prompts.β2,890Updated last year
- Efficient, scalable and enterprise-grade CPU/GPU inference server for π€ Hugging Face transformer models πβ1,686Updated 8 months ago
- A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)β4,672Updated last year
- Explain, analyze, and visualize NLP language models. Ecco creates interactive visualizations directly in Jupyter notebooks explaining theβ¦β2,040Updated 10 months ago
- π A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (iβ¦β8,875Updated this week
- Lighteval is your all-in-one toolkit for evaluating LLMs across multiple backendsβ1,670Updated this week
- General technology for enabling AI capabilities w/ LLMs and MLLMsβ4,037Updated 2 weeks ago
- Beyond the Imitation Game collaborative benchmark for measuring and extrapolating the capabilities of language modelsβ3,076Updated 11 months ago
- β1,226Updated 11 months ago
- Ongoing research training transformer language models at scale, including: BERT & GPT-2β1,397Updated last year
- Foundation Architecture for (M)LLMsβ3,081Updated last year
- SGPT: GPT Sentence Embeddings for Semantic Searchβ868Updated last year
- Pyserini is a Python toolkit for reproducible information retrieval research with sparse and dense representations.β1,879Updated this week
- Dense Passage Retriever - is a set of tools and models for open domain Q&A task.β1,813Updated 2 years ago
- MTEB: Massive Text Embedding Benchmarkβ2,648Updated this week
- Organize your experiments into discrete steps that can be cached and reused throughout the lifetime of your research project.β561Updated last year
- The implementation of DeBERTaβ2,108Updated last year
- Expanding natural instructionsβ1,005Updated last year