finetunej / gpt-neo_finetune_2.7B
β28Updated last year
Related projects: β
- Hidden Engrams: Long Term Memory for Transformer Model Inferenceβ34Updated 3 years ago
- π€Transformers: State-of-the-art Natural Language Processing for Pytorch and TensorFlow 2.0.β55Updated 2 years ago
- A basic ui for running gpt neo 2.7B on low vram (3 gb Vram minimum)β35Updated 3 years ago
- A ready-to-deploy container for implementing an easy to use REST API to access Language Models.β64Updated last year
- Conversational Language model toolkit for training against human preferences.β41Updated 5 months ago
- Colab notebooks to run a basic AI Dungeon clone using gpt-neo-2.7Bβ64Updated 3 years ago
- β9Updated 3 years ago
- Create soft prompts for fairseq 13B dense, GPT-J-6B and GPT-Neo-2.7B for free in a Google Colab TPU instanceβ27Updated last year
- A package for fine-tuning Transformers with TPUs, written in Tensorflow2.0+β37Updated 3 years ago
- Tools with GUI for GPT finetune data preparationβ23Updated 3 years ago
- GPT2Explorer is bringing GPT2 OpenAI langage models playground to run locally on standard windows computers.β29Updated 2 years ago
- β31Updated last year
- β110Updated last year
- Prompt tuning toolkit for GPT-2 and GPT-Neoβ88Updated 2 years ago
- Fork of kingoflolz/mesh-transformer-jax with memory usage optimizations and support for GPT-Neo, GPT-NeoX, BLOOM, OPT and fairseq dense Lβ¦β22Updated last year
- Dalle serviceβ50Updated 2 years ago
- RWKV-v2-RNN trained on the Pile. See https://github.com/BlinkDL/RWKV-LM for details.β65Updated 2 years ago
- β86Updated 2 years ago
- Experiments with generating opensource language model assistantsβ97Updated last year
- A notebook that runs GPT-Neo with low vram (6 gb) and cuda acceleration by loading it into gpu memory in smaller parts.β14Updated 3 years ago
- A search engine for ParlAI's BlenderBot project (and probably other ones as well)β132Updated 2 years ago
- Simple Annotated implementation of GPT-NeoX in PyTorchβ110Updated 2 years ago
- Fine-tuning 6-Billion GPT-J (& other models) with LoRA and 8-bit compressionβ65Updated last year
- β32Updated last year
- A repository to run gpt-j-6b on low vram machines (4.2 gb minimum vram for 2000 token context, 3.5 gb for 1000 token context). Model loadβ¦β114Updated 2 years ago
- β127Updated 2 years ago
- β50Updated last year
- Platform and API Agnostic library for powering chatbotsβ24Updated last year
- Tutorial to pretrain & fine-tune a π€ Flax T5 model on a TPUv3-8 with GCPβ58Updated 2 years ago
- β33Updated 3 years ago