yurakuratov / t5-experiments
Tools and scripts for experimenting with Transformers: Bert, T5...
☆54Updated last year
Alternatives and similar repositories for t5-experiments:
Users that are interested in t5-experiments are comparing it to the libraries listed below
- Helper scripts and notes that were used while porting various nlp models☆45Updated 2 years ago
- SWIM-IR is a Synthetic Wikipedia-based Multilingual Information Retrieval training set with 28 million query-passage pairs spanning 33 la…☆45Updated last year
- ☆96Updated 2 years ago
- QAmeleon introduces synthetic multilingual QA data using PaLM, a 540B large language model. This dataset was generated by prompt tuning P…☆34Updated last year
- Embedding Recycling for Language models☆38Updated last year
- [ICML 24 NGSM workshop] Associative Recurrent Memory Transformer implementation and scripts for training and evaluating☆34Updated last week
- Techniques used to run BLOOM at inference in parallel☆37Updated 2 years ago
- Starbucks: Improved Training for 2D Matryoshka Embeddings☆17Updated 3 months ago
- A fast implementation of T5/UL2 in PyTorch using Flash Attention☆78Updated this week
- Framework for probing tasks☆25Updated 10 months ago
- Transformers at any scale☆41Updated last year
- Supercharge huggingface transformers with model parallelism.☆76Updated 3 months ago
- ☆48Updated last year
- The codebase for our ACL2023 paper: Did You Read the Instructions? Rethinking the Effectiveness of Task Definitions in Instruction Learni…☆29Updated last year
- Hugging Face RoBERTa with Flash Attention 2☆21Updated last year
- LTG-Bert☆29Updated last year
- ☆29Updated 11 months ago
- This is a new metric that can be used to evaluate faithfulness of text generated by LLMs. The work behind this repository can be found he…☆31Updated last year
- A library for computing diverse text characteristics and using them to analyze data sets and models with ease.☆40Updated 2 years ago
- ☆67Updated 2 years ago
- Repo for ICML23 "Why do Nearest Neighbor Language Models Work?"☆56Updated 2 years ago
- ☆45Updated last year
- Minimum Bayes Risk Decoding for Hugging Face Transformers☆56Updated 7 months ago
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.☆93Updated last year
- [ICLR 2023] PyTorch code of Summarization Programs: Interpretable Abstractive Summarization with Neural Modular Trees☆23Updated last year
- Our open source implementation of MiniLMv2 (https://aclanthology.org/2021.findings-acl.188)☆60Updated last year
- Plug-and-play Search Interfaces with Pyserini and Hugging Face☆32Updated last year
- ☆45Updated 2 years ago
- Reference implementation for Reward-Augmented Decoding: Efficient Controlled Text Generation With a Unidirectional Reward Model☆42Updated last year
- The official repository for Efficient Long-Text Understanding Using Short-Text Models (Ivgi et al., 2022) paper☆68Updated last year