affjljoo3581 / GPT2
PyTorch Implementation of OpenAI GPT-2
☆334Updated 10 months ago
Alternatives and similar repositories for GPT2:
Users that are interested in GPT2 are comparing it to the libraries listed below
- Simple Text-Generator with OpenAI gpt-2 Pytorch Implementation☆992Updated 5 years ago
- Plain pytorch implementation of LLaMA☆187Updated last year
- Ongoing research training transformer language models at scale, including: BERT & GPT-2☆1,386Updated last year
- The universal integrated corpus-building environment.☆29Updated 4 years ago
- ☆345Updated 3 years ago
- Fine tune a T5 transformer model using PyTorch & Transformers🤗☆212Updated 4 years ago
- Code for the paper Fine-Tuning Language Models from Human Preferences☆1,323Updated last year
- Parallelformers: An Efficient Model Parallelization Toolkit for Deployment☆786Updated 2 years ago
- Prefix-Tuning: Optimizing Continuous Prompts for Generation☆923Updated last year
- Code for the ALiBi method for transformer language models (ICLR 2022)☆524Updated last year
- Implementation of RETRO, Deepmind's Retrieval based Attention net, in Pytorch☆864Updated last year
- A fast MoE impl for PyTorch☆1,713Updated 2 months ago
- A novel method to tune language models. Codes and datasets for paper ``GPT understands, too''.☆931Updated 2 years ago
- PyTorch Implementation of OpenAI GPT☆126Updated last year
- Implementation of "The Power of Scale for Parameter-Efficient Prompt Tuning"☆167Updated 3 years ago
- Ongoing research training transformer language models at scale, including: BERT & GPT-2☆2,062Updated last month
- A modular RL library to fine-tune language models to human preferences☆2,305Updated last year
- Tutorial for how to build BERT from scratch☆92Updated 11 months ago
- train llama on a single A100 80G node using 🤗 transformers and 🚀 Deepspeed Pipeline Parallelism☆219Updated last year
- Fast Inference Solutions for BLOOM☆561Updated 6 months ago
- This PyTorch package implements MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided Adaptation (NAACL 2022).☆104Updated 3 years ago
- A Pytorch implementation of Sparsely-Gated Mixture of Experts, for massively increasing the parameter count of language models☆736Updated last year
- Plug and Play Language Model implementation. Allows to steer topic and attributes of GPT-2 models.☆1,144Updated last year
- Automatically split your PyTorch models on multiple GPUs for training & inference☆651Updated last year
- This is a repository with the code for the ACL 2019 paper "Analyzing Multi-Head Self-Attention: Specialized Heads Do the Heavy Lifting, t…☆312Updated 3 years ago
- Scripts for fine-tuning Llama2 via SFT and DPO.☆197Updated last year
- Central place for the engineering/scaling WG: documentation, SLURM scripts and logs, compute environment and data.☆993Updated 9 months ago
- Transformer implementation in PyTorch.☆486Updated 6 years ago
- Few-shot Learning of GPT-3☆350Updated last year
- ☆526Updated last year