iliaschalkidis / flash-roberta
Hugging Face RoBERTa with Flash Attention 2
☆20Updated last year
Alternatives and similar repositories for flash-roberta:
Users that are interested in flash-roberta are comparing it to the libraries listed below
- Observe the slow deterioration of my mental sanity in the github commit history☆13Updated last year
- SWIM-IR is a Synthetic Wikipedia-based Multilingual Information Retrieval training set with 28 million query-passage pairs spanning 33 la…☆45Updated last year
- Embedding Recycling for Language models☆38Updated last year
- Starbucks: Improved Training for 2D Matryoshka Embeddings☆17Updated 3 months ago
- ☆21Updated 3 years ago
- ☆29Updated 11 months ago
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.☆93Updated last year
- ☆55Updated 2 years ago
- Dense hybrid representations for text retrieval☆61Updated last year
- No Parameter Left Behind: How Distillation and Model Size Affect Zero-Shot Retrieval☆28Updated 2 years ago
- A package for fine tuning of pretrained NLP transformers using Semi Supervised Learning☆15Updated 3 years ago
- ☆16Updated last year
- Pre-train Static Word Embeddings☆34Updated this week
- INCOME: An Easy Repository for Training and Evaluation of Index Compression Methods in Dense Retrieval. Includes BPR and JPQ.☆22Updated last year
- ☆46Updated 2 years ago
- [ACL 2023] Few-shot Reranking for Multi-hop QA via Language Model Prompting☆27Updated last year
- Tutorial to pretrain & fine-tune a 🤗 Flax T5 model on a TPUv3-8 with GCP☆58Updated 2 years ago
- Transformers at any scale☆41Updated last year
- DQ-BART: Efficient Sequence-to-Sequence Model via Joint Distillation and Quantization (ACL 2022)☆50Updated last year
- [ICLR 2023] PyTorch code of Summarization Programs: Interpretable Abstractive Summarization with Neural Modular Trees☆23Updated last year
- ☆28Updated last year
- My NER Experiments with ModernBERT☆15Updated last week
- Implementation of the paper 'Sentence Bottleneck Autoencoders from Transformer Language Models'☆17Updated 2 years ago
- A Benchmark for Robust, Multi-evidence, Multi-answer Question Answering☆16Updated 2 years ago
- ☆16Updated 5 months ago
- Plug-and-play Search Interfaces with Pyserini and Hugging Face☆32Updated last year
- QAmeleon introduces synthetic multilingual QA data using PaLM, a 540B large language model. This dataset was generated by prompt tuning P…☆34Updated last year
- Repo for ICML23 "Why do Nearest Neighbor Language Models Work?"☆56Updated 2 years ago
- Using short models to classify long texts☆21Updated last year