karpathy / minGPTLinks
A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training
β22,108Updated 10 months ago
Alternatives and similar repositories for minGPT
Users that are interested in minGPT are comparing it to the libraries listed below
Sorting:
- The simplest, fastest repository for training/finetuning medium-sized GPTs.β41,856Updated 6 months ago
- π€ PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.β18,774Updated last week
- Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"β12,090Updated 6 months ago
- QLoRA: Efficient Finetuning of Quantized LLMsβ10,500Updated last year
- Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalitiesβ21,420Updated 2 weeks ago
- A tiny scalar-valued autograd engine and a neural net library on top of it with PyTorch-like APIβ12,122Updated 10 months ago
- [NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.β22,795Updated 10 months ago
- An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.β38,731Updated 2 weeks ago
- Facebook AI Research Sequence-to-Sequence Toolkit written in Python.β31,543Updated last week
- Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adβ¦β6,073Updated 9 months ago
- Train transformer language models with reinforcement learning.β14,193Updated this week
- π A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (iβ¦β8,839Updated this week
- Fast and memory-efficient exact attentionβ17,846Updated this week
- A playbook for systematically maximizing the performance of deep learning models.β28,827Updated last year
- Instruct-tune LLaMA on consumer hardwareβ18,917Updated 10 months ago
- A collection of libraries to optimise AI model performancesβ8,373Updated 10 months ago
- The fastai deep learning libraryβ27,064Updated 2 weeks ago
- Unsupervised text tokenizer for Neural Network-based text generation.β10,994Updated 2 months ago
- π€ Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal modelβ¦β145,689Updated this week
- A curated list of practical guide resources of LLMs (LLMs Tree, Examples, Papers)β9,947Updated last year
- DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.β38,997Updated this week
- Neural Networks: Zero to Heroβ14,036Updated 10 months ago
- Tensor library for machine learningβ12,697Updated last week
- RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable)β¦β13,706Updated last week
- Code and documentation to train Stanford's Alpaca models, and generate the data.β30,040Updated 11 months ago
- tiktoken is a fast BPE tokeniser for use with OpenAI's models.β14,829Updated 3 months ago
- State-of-the-Art Text Embeddingsβ16,947Updated last week
- Ongoing research training transformer models at scaleβ12,600Updated this week
- Inference Llama 2 in one file of pure Cβ18,475Updated 10 months ago
- Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.β9,700Updated 11 months ago