affjljoo3581 / GPT2
PyTorch Implementation of OpenAI GPT-2
☆288Updated 4 months ago
Related projects ⓘ
Alternatives and complementary repositories for GPT2
- Simple Text-Generator with OpenAI gpt-2 Pytorch Implementation☆972Updated 5 years ago
- A novel method to tune language models. Codes and datasets for paper ``GPT understands, too''.☆923Updated 2 years ago
- Code for the ALiBi method for transformer language models (ICLR 2022)☆506Updated last year
- Diffusion-LM☆1,055Updated 3 months ago
- Prefix-Tuning: Optimizing Continuous Prompts for Generation☆895Updated 6 months ago
- [ICLR'23] DiffuSeq: Sequence to Sequence Text Generation with Diffusion Models☆731Updated 8 months ago
- PyTorch Implementation of OpenAI GPT☆118Updated last year
- Transformer implementation in PyTorch.☆464Updated 5 years ago
- Pytorch Implementation of Google BERT☆590Updated 4 years ago
- Rotary Transformer☆812Updated 2 years ago
- A research project for natural language generation, containing the official implementations by MSRA NLC team.☆690Updated 3 months ago
- [NeurIPS'22 Spotlight] A Contrastive Framework for Neural Text Generation☆466Updated 8 months ago
- Fine tune a T5 transformer model using PyTorch & Transformers🤗☆196Updated 3 years ago
- Ongoing research training transformer language models at scale, including: BERT & GPT-2☆1,335Updated 7 months ago
- Optimus: the first large-scale pre-trained VAE language model☆369Updated last year
- ☆484Updated last year
- A PyTorch implementation of Transformer in "Attention is All You Need"☆103Updated 3 years ago
- Open-source pre-training implementation of Google's LaMDA in PyTorch. Adding RLHF similar to ChatGPT.☆466Updated 8 months ago
- A simple and working implementation of Electra, the fastest way to pretrain language models from scratch, in Pytorch☆222Updated last year
- ☆339Updated 3 years ago
- An implementation of masked language modeling for Pytorch, made as concise and simple as possible☆177Updated last year
- Automatically split your PyTorch models on multiple GPUs for training & inference☆624Updated 10 months ago
- Longformer: The Long-Document Transformer☆2,045Updated last year
- Code for the paper Fine-Tuning Language Models from Human Preferences☆1,227Updated last year
- Dense Passage Retriever - is a set of tools and models for open domain Q&A task.☆1,723Updated last year
- Guide: Finetune GPT2-XL (1.5 Billion Parameters) and finetune GPT-NEO (2.7 B) on a single GPU with Huggingface Transformers using DeepSpe…☆432Updated last year
- Expanding natural instructions☆956Updated 11 months ago
- A Pytorch implementation of Sparsely-Gated Mixture of Experts, for massively increasing the parameter count of language models☆639Updated last year
- train llama on a single A100 80G node using 🤗 transformers and 🚀 Deepspeed Pipeline Parallelism☆207Updated 11 months ago
- Few-shot Learning of GPT-3☆342Updated last year