openai / gpt-2
Code for the paper "Language Models are Unsupervised Multitask Learners"
β23,310Updated 8 months ago
Alternatives and similar repositories for gpt-2:
Users that are interested in gpt-2 are comparing it to the libraries listed below
- GPT-3: Language Models are Few-Shot Learnersβ15,759Updated 4 years ago
- π€ Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.β142,871Updated this week
- Facebook AI Research Sequence-to-Sequence Toolkit written in Python.β31,297Updated 3 months ago
- Dataset of GPT-2 outputs for research in detection, biases, and moreβ1,972Updated last year
- An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.β8,288Updated 3 years ago
- π₯ Fast State-of-the-Art Tokenizers optimized for Research and Productionβ9,592Updated 3 weeks ago
- DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.β37,916Updated this week
- TensorFlow code and pre-trained models for BERTβ39,027Updated 8 months ago
- π€ The largest hub of ready-to-use datasets for ML models with fast, easy-to-use and efficient data manipulation toolsβ19,969Updated last week
- An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed librariesβ7,157Updated this week
- Unsupervised text tokenizer for Neural Network-based text generation.β10,786Updated 2 weeks ago
- BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)β7,306Updated last year
- Python package to easily retrain OpenAI's GPT-2 text-generating model on new textsβ3,404Updated 2 years ago
- Ongoing research training transformer models at scaleβ12,075Updated this week
- Open Source Neural Machine Translation and (Large) Language Models in PyTorchβ6,863Updated last month
- CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an imageβ28,419Updated 8 months ago
- A modular framework for vision & language multimodal research from Facebook AI Research (FAIR)β5,557Updated last week
- State-of-the-Art Deep Learning scripts organized by models - easy to train and deploy with reproducible accuracy and performance on enterβ¦β14,158Updated 8 months ago
- XLNet: Generalized Autoregressive Pretraining for Language Understandingβ6,187Updated last year
- A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Autoβ¦β13,658Updated this week
- Code and model for the paper "Improving Language Understanding by Generative Pre-Training"β2,202Updated 6 years ago
- Build and share delightful machine learning apps, all in Python. π Star to support our work!β37,426Updated this week
- Pretrain, finetune ANY AI model of ANY size on multiple GPUs, TPUs with zero code changes.β29,300Updated this week
- An annotated implementation of the Transformer paper.β6,157Updated last year
- Tensors and Dynamic neural networks in Python with strong GPU accelerationβ89,003Updated this week
- π A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (iβ¦β8,608Updated this week
- Code for the paper "Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer"β6,329Updated last month
- Library of deep learning models and datasets designed to make deep learning more accessible and accelerate ML research.β16,038Updated last year
- Official implementations for various pre-training models of ERNIE-family, covering topics of Language Understanding & Generation, Multimoβ¦β6,369Updated 7 months ago
- Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalitiesβ21,069Updated last month