AI21Labs / lm-evaluation
Evaluation suite for large-scale language models.
☆124Updated 3 years ago
Alternatives and similar repositories for lm-evaluation:
Users that are interested in lm-evaluation are comparing it to the libraries listed below
- ☆77Updated last year
- A diff tool for language models☆42Updated last year
- ☆66Updated 2 years ago
- A framework for few-shot evaluation of autoregressive language models.☆103Updated last year
- Pipeline for pulling and processing online language model pretraining data from the web☆177Updated last year
- Apps built using Inspired Cognition's Critique.☆58Updated 2 years ago
- Implementation of Marge, Pre-training via Paraphrasing, in Pytorch☆75Updated 4 years ago
- A library to create and manage configuration files, especially for machine learning projects.☆77Updated 3 years ago
- Experiments with generating opensource language model assistants☆97Updated last year
- ☆89Updated 2 years ago
- Used for adaptive human in the loop evaluation of language and embedding models.☆307Updated 2 years ago
- Our open source implementation of MiniLMv2 (https://aclanthology.org/2021.findings-acl.188)☆61Updated last year
- Library for soft prompt tuning☆23Updated last year
- Open source library for few shot NLP☆78Updated last year
- ☆97Updated 2 years ago
- Python tools for processing the stackexchange data dumps into a text dataset for Language Models☆81Updated last year
- Transformers at any scale☆41Updated last year
- ☆44Updated 4 months ago
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.☆93Updated 2 years ago
- ☆147Updated 4 years ago
- ☆132Updated last year
- ☆72Updated last year
- For experiments involving instruct gpt. Currently used for documenting open research questions.☆71Updated 2 years ago
- ☆93Updated 3 months ago
- 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.☆82Updated 3 years ago
- Code and data to support the paper "PAQ 65 Million Probably-Asked Questions andWhat You Can Do With Them"☆202Updated 3 years ago
- Techniques used to run BLOOM at inference in parallel☆37Updated 2 years ago
- AuditNLG: Auditing Generative AI Language Modeling for Trustworthiness☆99Updated 2 months ago
- This project studies the performance and robustness of language models and task-adaptation methods.☆149Updated 10 months ago
- Automatic metrics for GEM tasks☆65Updated 2 years ago