allenai / tango
Organize your experiments into discrete steps that can be cached and reused throughout the lifetime of your research project.
β554Updated 11 months ago
Alternatives and similar repositories for tango:
Users that are interested in tango are comparing it to the libraries listed below
- π€ A PyTorch library of curated Transformer models and their composable componentsβ885Updated last year
- Mistral: A strong, northwesterly wind: Framework for transparent and accessible large-scale language model training, built with Hugging Fβ¦β572Updated last year
- An open collection of implementation tips, tricks and resources for training large language modelsβ472Updated 2 years ago
- π€ Evaluate: A library for easily evaluating machine learning models and datasets.β2,203Updated 3 months ago
- Task-based datasets, preprocessing, and evaluation for sequence models.β574Updated last week
- Flexible components pairing π€ Transformers with Pytorch Lightningβ608Updated 2 years ago
- Cramming the training of a (BERT-type) language model into limited compute.β1,331Updated 10 months ago
- Fast & Simple repository for pre-training and fine-tuning T5-style modelsβ1,003Updated 8 months ago
- Run Effective Large Batch Contrastive Learning Beyond GPU/TPU Memory Constraintβ386Updated last year
- Reproduce results and replicate training fo T0 (Multitask Prompted Training Enables Zero-Shot Task Generalization)β463Updated 2 years ago
- Build, evaluate, understand, and fix LLM-based appsβ488Updated last year
- AI Data Management & Evaluation Platformβ215Updated last year
- String-to-String Algorithms for Natural Language Processingβ545Updated 9 months ago
- Interpretability for sequence generation models π πβ413Updated 2 weeks ago
- Implementation of RETRO, Deepmind's Retrieval based Attention net, in Pytorchβ864Updated last year
- Central place for the engineering/scaling WG: documentation, SLURM scripts and logs, compute environment and data.β993Updated 9 months ago
- Contriever: Unsupervised Dense Information Retrieval with Contrastive Learningβ732Updated 2 years ago
- Interpretable Evaluation for AI Systemsβ366Updated 2 years ago
- Powerful unsupervised domain adaptation method for dense retrieval. Requires only unlabeled corpus and yields massive improvement: "GPL: β¦β333Updated last year
- maximal update parametrization (Β΅P)β1,500Updated 9 months ago
- Repository containing code for "How to Train BERT with an Academic Budget" paperβ313Updated last year
- Code repository for supporting the paper "Atlas Few-shot Learning with Retrieval Augmented Language Models",(https//arxiv.org/abs/2208.03β¦β535Updated last year
- Original Implementation of Prompt Tuning from Lester, et al, 2021β679Updated 2 months ago
- Code for the ALiBi method for transformer language models (ICLR 2022)β524Updated last year
- NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Dayβ255Updated last year
- Creative interactive views of any dataset.β838Updated 4 months ago
- Recipes are a standard, well supported set of blueprints for machine learning engineers to rapidly train models using the latest researchβ¦β312Updated this week
- Datasets collection and preprocessings framework for NLP extreme multitask learningβ180Updated 4 months ago
- Kernl lets you run PyTorch transformer models several times faster on GPU with a single line of code, and is designed to be easily hackabβ¦β1,566Updated last year
- β505Updated last year