openai / gpt-2Links
Code for the paper "Language Models are Unsupervised Multitask Learners"
☆23,680Updated 10 months ago
Alternatives and similar repositories for gpt-2
Users that are interested in gpt-2 are comparing it to the libraries listed below
Sorting:
- GPT-3: Language Models are Few-Shot Learners☆15,753Updated 4 years ago
- Unsupervised text tokenizer for Neural Network-based text generation.☆10,994Updated 2 months ago
- A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training☆22,108Updated 10 months ago
- TensorFlow code and pre-trained models for BERT☆39,244Updated 10 months ago
- Dataset of GPT-2 outputs for research in detection, biases, and more☆1,983Updated last year
- 🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal model…☆145,689Updated this week
- Python package to easily retrain OpenAI's GPT-2 text-generating model on new texts☆3,406Updated 2 years ago
- Software in C and data files for the popular GloVe model for distributed word representations, a.k.a. word vectors or embeddings☆7,047Updated 6 months ago
- Facebook AI Research Sequence-to-Sequence Toolkit written in Python.☆31,543Updated last week
- Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities☆21,420Updated 2 weeks ago
- Detectron2 is a platform for object detection, segmentation and other visual recognition tasks.☆32,197Updated this week
- XLNet: Generalized Autoregressive Pretraining for Language Understanding☆6,183Updated 2 years ago
- Ongoing research training transformer models at scale☆12,600Updated this week
- Code for the paper "Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer"☆6,379Updated last month
- 🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.☆18,820Updated this week
- DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.☆38,997Updated this week
- Inference code for Llama models☆58,399Updated 4 months ago
- Code and model for the paper "Improving Language Understanding by Generative Pre-Training"☆2,214Updated 6 years ago
- Google AI 2018 BERT pytorch implementation☆6,419Updated last year
- Development repository for the Triton language and compiler☆15,881Updated this week
- Code and documentation to train Stanford's Alpaca models, and generate the data.☆30,040Updated 11 months ago
- An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries☆7,228Updated last week
- Google Research☆35,790Updated this week
- Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"☆12,090Updated 6 months ago
- Code for the paper "Language Models are Unsupervised Multitask Learners"☆1,146Updated 2 years ago
- tiktoken is a fast BPE tokeniser for use with OpenAI's models.☆14,829Updated 3 months ago
- Fast and memory-efficient exact attention☆17,952Updated this week
- ☆3,658Updated 2 years ago
- Library of deep learning models and datasets designed to make deep learning more accessible and accelerate ML research.☆16,218Updated 2 years ago
- ALBERT: A Lite BERT for Self-supervised Learning of Language Representations☆3,272Updated 2 years ago