mkturkcan / GPTune
Fine tuning experiments for the GPT-2 model by OpenAI.
☆20Updated 5 years ago
Alternatives and similar repositories for GPTune:
Users that are interested in GPTune are comparing it to the libraries listed below
- GPT-2 User Interface based on HuggingFace's Pytorch Implementation☆56Updated 8 months ago
- Fine tune GPT-2 with your favourite authors☆72Updated last year
- Code for Improving Neural Story Generation by Targeted Common Sense Grounding☆41Updated 5 years ago
- Experiments with generating GPT-2 fanfiction on specified topics.☆11Updated 5 years ago
- Agents that build knowledge graphs and explore textual worlds by asking questions☆79Updated last year
- The elegant integration of huggingface/nlp and fastai2 and handy transforms using pure huggingface/nlp☆19Updated 4 years ago
- Code for the paper "Language Models are Unsupervised Multitask Learners"☆109Updated 3 years ago
- Code for my blog post on Generating Words from Embeddings☆23Updated 8 months ago
- ☆19Updated 2 years ago
- A Flask Web App for Generating Text with GPT-2☆37Updated 6 years ago
- Hidden Engrams: Long Term Memory for Transformer Model Inference☆35Updated 3 years ago
- The projects lets you extract glossary words and their definitions from a given piece of text automatically using NLP techniques☆29Updated 4 years ago
- A package for fine-tuning Transformers with TPUs, written in Tensorflow2.0+☆37Updated 4 years ago
- A set of tools for leveraging pre-trained embeddings, active learning and model explainability for effecient document classification☆29Updated 2 months ago
- Reddit title generator API based on GPT-2☆19Updated 5 years ago
- interactive explorer for language models☆133Updated 3 years ago
- Given a pair of sentences (premise, hypothesis), the decomposed graph entailment model (DGEM) predicts whether the premise can be used to…☆52Updated 4 years ago
- Steering Vector Repo from "Extracting Latent Steering Vectors from Pretrained Language Models" - ACL2022 Findings☆10Updated 3 years ago
- Create soft prompts for fairseq 13B dense, GPT-J-6B and GPT-Neo-2.7B for free in a Google Colab TPU instance