VatsaDev / nanoChatGPTLinks
nanogpt turned into a chat model
☆76Updated 2 years ago
Alternatives and similar repositories for nanoChatGPT
Users that are interested in nanoChatGPT are comparing it to the libraries listed below
Sorting:
- Micro Llama is a small Llama based model with 300M parameters trained from scratch with $500 budget☆161Updated 2 months ago
 - Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆231Updated last year
 - Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first app…☆169Updated last year
 - an implementation of Self-Extend, to expand the context window via grouped attention☆118Updated last year
 - ☆51Updated last year
 - Low-Rank adapter extraction for fine-tuned transformers models☆178Updated last year
 - Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆78Updated last year
 - Implementation of the Mamba SSM with hf_integration.☆56Updated last year
 - ☆88Updated last year
 - Scripts to create your own moe models using mlx☆90Updated last year
 - Some simple scripts that I use day-to-day when working with LLMs and Huggingface Hub☆160Updated 2 years ago
 - Finetune Falcon, LLaMA, MPT, and RedPajama on consumer hardware using PEFT LoRA☆103Updated 5 months ago
 - Small and Efficient Mathematical Reasoning LLMs☆72Updated last year
 - Collection of autoregressive model implementation☆86Updated 6 months ago
 - Video+code lecture on building nanoGPT from scratch☆68Updated last year
 - Train your own small bitnet model☆75Updated last year
 - Tune MPTs☆84Updated 2 years ago
 - Convenient wrapper for fine-tuning and inference of Large Language Models (LLMs) with several quantization techniques (GTPQ, bitsandbytes…☆146Updated 2 years ago
 - ☆78Updated last year
 - The simplest, fastest repository for training/finetuning medium-sized xLSTMs.☆41Updated last year
 - minimal scripts for 24GB VRAM GPUs. training, inference, whatever☆43Updated this week
 - Data preparation code for Amber 7B LLM☆93Updated last year
 - ☆95Updated 2 years ago
 - Pre-training code for Amber 7B LLM☆169Updated last year
 - This repo is for handling Question Answering, especially for Multi-hop Question Answering☆67Updated last year
 - Load multiple LoRA modules simultaneously and automatically switch the appropriate combination of LoRA modules to generate the best answe…☆158Updated last year
 - ☆50Updated 8 months ago
 - This project is a collection of fine-tuning scripts to help researchers fine-tune Qwen 2 VL on HuggingFace datasets.☆74Updated 3 months ago
 - GPT-2 small trained on phi-like data☆67Updated last year
 - Using open source LLMs to build synthetic datasets for direct preference optimization☆68Updated last year