google-research / t5x
☆2,767Updated 2 weeks ago
Alternatives and similar repositories for t5x:
Users that are interested in t5x are comparing it to the libraries listed below
- The hub for EleutherAI's work on interpretability and learning dynamics☆2,413Updated last week
- A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)☆4,601Updated last year
- Accessible large language models via k-bit quantization for PyTorch.☆6,818Updated this week
- The RedPajama-Data repository contains code for preparing large datasets for training large language models.☆4,678Updated 3 months ago
- Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Fl…☆2,459Updated 7 months ago
- Ongoing research training transformer language models at scale, including: BERT & GPT-2☆1,378Updated last year
- PyTorch extensions for high performance and large scale training.☆3,278Updated 2 months ago
- A modular RL library to fine-tune language models to human preferences☆2,288Updated last year
- Toolkit for creating, sharing and using natural language prompts.☆2,798Updated last year
- Implementation of Toolformer, Language Models That Can Use Tools, by MetaAI☆2,013Updated 7 months ago
- ☆1,507Updated last week
- Beyond the Imitation Game collaborative benchmark for measuring and extrapolating the capabilities of language models☆2,992Updated 8 months ago
- Central place for the engineering/scaling WG: documentation, SLURM scripts and logs, compute environment and data.☆988Updated 7 months ago
- Holistic Evaluation of Language Models (HELM), a framework to increase the transparency of language models (https://arxiv.org/abs/2211.09…☆2,116Updated this week
- Human preference data for "Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback"☆1,703Updated last year
- maximal update parametrization (µP)☆1,480Updated 8 months ago
- An open-source framework for training large multimodal models.☆3,845Updated 6 months ago
- Code for the paper "Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer"☆6,298Updated 3 weeks ago
- Fast & Simple repository for pre-training and fine-tuning T5-style models☆999Updated 7 months ago
- 🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization…☆2,812Updated 2 weeks ago
- Aligning pretrained language models with instruction data generated by themselves.☆4,314Updated last year
- [ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters☆5,832Updated last year
- 🤗 Evaluate: A library for easily evaluating machine learning models and datasets.☆2,152Updated 2 months ago
- QLoRA: Efficient Finetuning of Quantized LLMs☆10,322Updated 9 months ago
- Training and serving large-scale neural networks with auto parallelization.☆3,114Updated last year
- Cramming the training of a (BERT-type) language model into limited compute.☆1,324Updated 9 months ago
- Transformer related optimization, including BERT, GPT☆6,084Updated 11 months ago
- Implementation of RETRO, Deepmind's Retrieval based Attention net, in Pytorch☆859Updated last year
- MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.☆1,987Updated last week
- Ongoing research training transformer language models at scale, including: BERT & GPT-2☆2,022Updated 3 weeks ago