allenai / tangoLinks
Organize your experiments into discrete steps that can be cached and reused throughout the lifetime of your research project.
โ561Updated last year
Alternatives and similar repositories for tango
Users that are interested in tango are comparing it to the libraries listed below
Sorting:
- Task-based datasets, preprocessing, and evaluation for sequence models.โ574Updated 3 weeks ago
- Flexible components pairing ๐ค Transformers with Pytorch Lightningโ609Updated 2 years ago
- An open collection of implementation tips, tricks and resources for training large language modelsโ473Updated 2 years ago
- Reproduce results and replicate training fo T0 (Multitask Prompted Training Enables Zero-Shot Task Generalization)โ463Updated 2 years ago
- Interpretable Evaluation for AI Systemsโ367Updated 2 years ago
- ๐ค A PyTorch library of curated Transformer models and their composable componentsโ888Updated last year
- Run Effective Large Batch Contrastive Learning Beyond GPU/TPU Memory Constraintโ389Updated last year
- Repository containing code for "How to Train BERT with an Academic Budget" paperโ313Updated last year
- Fast & Simple repository for pre-training and fine-tuning T5-style modelsโ1,004Updated 9 months ago
- Mistral: A strong, northwesterly wind: Framework for transparent and accessible large-scale language model training, built with Hugging Fโฆโ572Updated last year
- Code for T-Few from "Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning"โ451Updated last year
- Interpretability for sequence generation models ๐ ๐โ419Updated last month
- Cramming the training of a (BERT-type) language model into limited compute.โ1,332Updated 11 months ago
- โ507Updated last year
- A data set based on all arXiv publications, pre-processed for NLP, including structured full-text and citation networkโ289Updated 8 months ago
- Code for the ALiBi method for transformer language models (ICLR 2022)โ530Updated last year
- All-in-one text de-duplicationโ679Updated last week
- Central place for the engineering/scaling WG: documentation, SLURM scripts and logs, compute environment and data.โ997Updated 10 months ago
- โ1,218Updated 10 months ago
- Model explainability that works seamlessly with ๐ค transformers. Explain your transformers model in just 2 lines of code.โ1,346Updated last year
- Code repository for the paper - "Matryoshka Representation Learning"โ497Updated last year
- BLEURT is a metric for Natural Language Generation based on transfer learning.โ733Updated last year
- Library for 8-bit optimizers and quantization routines.โ716Updated 2 years ago
- W&B Server is the self hosted version of Weights & Biasesโ278Updated this week
- A Visual Analysis Tool to Explore Learned Representations in Transformers Modelsโ591Updated last year
- Powerful unsupervised domain adaptation method for dense retrieval. Requires only unlabeled corpus and yields massive improvement: "GPL: โฆโ333Updated last year
- git extension for {collaborative, communal, continual} model developmentโ212Updated 6 months ago
- NL-Augmenter ๐ฆ โ ๐ A Collaborative Repository of Natural Language Transformationsโ786Updated last year
- Officially supported AllenNLP modelsโ543Updated 2 years ago
- maximal update parametrization (ยตP)โ1,526Updated 10 months ago