huggingface / evaluateLinks
π€ Evaluate: A library for easily evaluating machine learning models and datasets.
β2,285Updated last month
Alternatives and similar repositories for evaluate
Users that are interested in evaluate are comparing it to the libraries listed below
Sorting:
- β1,532Updated this week
- A Unified Library for Parameter-Efficient and Modular Transfer Learningβ2,748Updated 2 months ago
- π Accelerate inference and training of π€ Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimizationβ¦β3,016Updated this week
- Beyond the Imitation Game collaborative benchmark for measuring and extrapolating the capabilities of language modelsβ3,097Updated last year
- Holistic Evaluation of Language Models (HELM) is an open source Python framework created by the Center for Research on Foundation Models β¦β2,405Updated this week
- A modular RL library to fine-tune language models to human preferencesβ2,333Updated last year
- Toolkit for creating, sharing and using natural language prompts.β2,917Updated last year
- β1,235Updated last year
- A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)β4,692Updated last year
- Efficient few-shot learning with Sentence Transformersβ2,544Updated last week
- Central place for the engineering/scaling WG: documentation, SLURM scripts and logs, compute environment and data.β1,007Updated last year
- PyTorch extensions for high performance and large scale training.β3,352Updated 3 months ago
- β2,858Updated 2 months ago
- The hub for EleutherAI's work on interpretability and learning dynamicsβ2,582Updated 2 months ago
- Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.β2,524Updated last week
- Human preference data for "Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback"β1,770Updated last month
- Efficient, scalable and enterprise-grade CPU/GPU inference server for π€ Hugging Face transformer models πβ1,687Updated 9 months ago
- Model explainability that works seamlessly with π€ transformers. Explain your transformers model in just 2 lines of code.β1,364Updated last year
- Cramming the training of a (BERT-type) language model into limited compute.β1,344Updated last year
- The implementation of DeBERTaβ2,131Updated last year
- General technology for enabling AI capabilities w/ LLMs and MLLMsβ4,082Updated last month
- Ongoing research training transformer language models at scale, including: BERT & GPT-2β1,406Updated last year
- π A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (iβ¦β9,010Updated this week
- Accessible large language models via k-bit quantization for PyTorch.β7,450Updated this week
- MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.β2,044Updated last month
- Original Implementation of Prompt Tuning from Lester, et al, 2021β690Updated 5 months ago
- Organize your experiments into discrete steps that can be cached and reused throughout the lifetime of your research project.β562Updated last year
- [ACL 2023] One Embedder, Any Task: Instruction-Finetuned Text Embeddingsβ1,995Updated 6 months ago
- Foundation Architecture for (M)LLMsβ3,101Updated last year
- Data and tools for generating and inspecting OLMo pre-training data.β1,290Updated this week