affjljoo3581 / GPT2
PyTorch Implementation of OpenAI GPT-2
☆312Updated 7 months ago
Alternatives and similar repositories for GPT2:
Users that are interested in GPT2 are comparing it to the libraries listed below
- PyTorch Implementation of OpenAI GPT☆121Updated last year
- Simple Text-Generator with OpenAI gpt-2 Pytorch Implementation☆985Updated 5 years ago
- Code for the ALiBi method for transformer language models (ICLR 2022)☆515Updated last year
- Ongoing research training transformer language models at scale, including: BERT & GPT-2☆1,364Updated 11 months ago
- An implementation of masked language modeling for Pytorch, made as concise and simple as possible☆178Updated last year
- The universal integrated corpus-building environment.☆29Updated 4 years ago
- Fine tune a T5 transformer model using PyTorch & Transformers🤗☆207Updated 4 years ago
- Implementation of "The Power of Scale for Parameter-Efficient Prompt Tuning"☆164Updated 3 years ago
- Fine-tuning GPT-2 Small for Question Answering☆129Updated 2 years ago
- Parallelformers: An Efficient Model Parallelization Toolkit for Deployment☆785Updated last year
- Guide: Finetune GPT2-XL (1.5 Billion Parameters) and finetune GPT-NEO (2.7 B) on a single GPU with Huggingface Transformers using DeepSpe…☆437Updated last year
- A novel method to tune language models. Codes and datasets for paper ``GPT understands, too''.☆928Updated 2 years ago
- Prefix-Tuning: Optimizing Continuous Prompts for Generation☆910Updated 9 months ago
- [NeurIPS'22 Spotlight] A Contrastive Framework for Neural Text Generation☆469Updated 11 months ago
- Official PyTorch implementation of DistiLLM: Towards Streamlined Distillation for Large Language Models (ICML 2024)☆186Updated 5 months ago
- Pytorch Implementation of Google BERT☆593Updated 4 years ago
- Code for "Learning to summarize from human feedback"☆1,006Updated last year
- Automatically split your PyTorch models on multiple GPUs for training & inference☆649Updated last year
- Diffusion-LM☆1,089Updated 6 months ago
- Code for T-Few from "Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning"☆443Updated last year
- This PyTorch package implements MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided Adaptation (NAACL 2022).☆100Updated 2 years ago
- Official code for ReLoRA from the paper Stack More Layers Differently: High-Rank Training Through Low-Rank Updates☆443Updated 10 months ago
- Tutorial for how to build BERT from scratch☆87Updated 9 months ago
- LLaMA 2 implemented from scratch in PyTorch☆294Updated last year
- The PyTorch implementation of fine-tuning the GPT-2(Generative Pre-trained Transformer 2) for dialogue generation.☆174Updated 8 months ago
- Code repository for supporting the paper "Atlas Few-shot Learning with Retrieval Augmented Language Models",(https//arxiv.org/abs/2208.03…☆526Updated last year
- A minimum example of aligning language models with RLHF similar to ChatGPT☆217Updated last year
- [ACL 2021] LM-BFF: Better Few-shot Fine-tuning of Language Models https://arxiv.org/abs/2012.15723☆725Updated 2 years ago
- Plain pytorch implementation of LLaMA☆189Updated last year
- ☆96Updated last year