embeddings-benchmark / arena
Code for the MTEB Arena
☆18Updated 5 months ago
Alternatives and similar repositories for arena:
Users that are interested in arena are comparing it to the libraries listed below
- minimal pytorch implementation of bm25 (with sparse tensors)☆97Updated last year
- ☆38Updated 10 months ago
- Datasets collection and preprocessings framework for NLP extreme multitask learning☆175Updated last month
- Code for Zero-Shot Tokenizer Transfer☆121Updated last month
- [Data + code] ExpertQA : Expert-Curated Questions and Attributed Answers☆125Updated 11 months ago
- ☆57Updated 5 months ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆99Updated 3 months ago
- Experiments with generating opensource language model assistants☆97Updated last year
- ☆73Updated 10 months ago
- Repository containing the SPIN experiments on the DIBT 10k ranked prompts☆24Updated 11 months ago
- Using open source LLMs to build synthetic datasets for direct preference optimization☆58Updated last year
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆72Updated 4 months ago
- Repo to hold code and track issues for the collection of permissively licensed data☆23Updated 2 months ago
- Exploring finetuning public checkpoints on filter 8K sequences on Pile☆115Updated last year
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment☆54Updated 6 months ago
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.☆82Updated last year
- ☆122Updated 2 weeks ago
- Manage scalable open LLM inference endpoints in Slurm clusters☆253Updated 7 months ago
- Fast, Modern, Memory Efficient, and Low Precision PyTorch Optimizers☆84Updated 7 months ago
- Minimum Bayes Risk Decoding for Hugging Face Transformers☆56Updated 9 months ago
- some common Huggingface transformers in maximal update parametrization (µP)☆79Updated 2 years ago
- Scaling is a distributed training library and installable dependency designed to scale up neural networks, with a dedicated module for tr…☆56Updated 4 months ago
- ☆73Updated last year
- Multipack distributed sampler for fast padding-free training of LLMs☆184Updated 6 months ago
- A fast implementation of T5/UL2 in PyTorch using Flash Attention☆82Updated last month
- ☆48Updated 3 months ago
- A collection of datasets for language model pretraining including scripts for downloading, preprocesssing, and sampling.☆56Updated 7 months ago
- Simple replication of [ColBERT-v1](https://arxiv.org/abs/2004.12832).☆80Updated 11 months ago
- experiments with inference on llama☆104Updated 8 months ago
- The Batched API provides a flexible and efficient way to process multiple requests in a batch, with a primary focus on dynamic batching o…☆123Updated 2 months ago