EleutherAI / gpt-neoLinks
An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
☆8,287Updated 3 years ago
Alternatives and similar repositories for gpt-neo
Users that are interested in gpt-neo are comparing it to the libraries listed below
Sorting:
- An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries☆7,188Updated 2 weeks ago
- Model parallel transformers in JAX and Haiku☆6,333Updated 2 years ago
- GPT-3: Language Models are Few-Shot Learners☆15,752Updated 4 years ago
- RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable)…☆13,645Updated this week
- Repo for external large-scale work☆6,527Updated last year
- Facebook AI Research Sequence-to-Sequence Toolkit written in Python.☆31,462Updated 4 months ago
- Implementation / replication of DALL-E, OpenAI's Text to Image Transformer, in Pytorch☆5,617Updated last year
- Code and documentation to train Stanford's Alpaca models, and generate the data.☆30,019Updated 10 months ago
- ☆9,010Updated last year
- ChatRWKV is like ChatGPT but powered by RWKV (100% RNN) language model, and open source.☆9,486Updated 3 weeks ago
- Running large language models on a single GPU for throughput-oriented scenarios.☆9,320Updated 7 months ago
- Unsupervised text tokenizer for Neural Network-based text generation.☆10,942Updated last month
- Instruct-tune LLaMA on consumer hardware☆18,904Updated 10 months ago
- A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)☆4,652Updated last year
- QLoRA: Efficient Finetuning of Quantized LLMs☆10,446Updated 11 months ago
- Locally run an Instruction-Tuned Chat-Style LLM☆10,229Updated 2 years ago
- ☆4,568Updated last year
- DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.☆38,566Updated this week
- StableLM: Stability AI Language Models☆15,832Updated last year
- Inference code for Llama models☆58,290Updated 4 months ago
- A framework for training and evaluating AI models on a variety of openly available dialogue datasets.☆10,531Updated last year
- Implementation of DALL-E 2, OpenAI's updated text-to-image synthesis neural network, in Pytorch☆11,277Updated last year
- The RedPajama-Data repository contains code for preparing large datasets for training large language models.☆4,727Updated 5 months ago
- A robust Python tool for text-based AI training and generation using GPT-2.☆1,842Updated last year
- The goal of this project is to enable users to create cool web demos using the newly released OpenAI GPT-3 API with just a few lines of P…☆2,893Updated last year
- Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architecture. Basically ChatGPT but with PaLM☆7,832Updated last month
- GLM-130B: An Open Bilingual Pre-Trained Model (ICLR 2023)☆7,682Updated last year
- 🏄 Scalable embedding, reasoning, ranking for images and sentences with CLIP☆12,672Updated last year
- 💥 Fast State-of-the-Art Tokenizers optimized for Research and Production☆9,726Updated this week
- Implementation of Imagen, Google's Text-to-Image Neural Network, in Pytorch☆8,268Updated 7 months ago