VatsaDev / nanoChatGPT
nanogpt turned into a chat model
☆65Updated last year
Alternatives and similar repositories for nanoChatGPT:
Users that are interested in nanoChatGPT are comparing it to the libraries listed below
- The simplest, fastest repository for training/finetuning medium-sized xLSTMs.☆39Updated 8 months ago
- Implementation of the Mamba SSM with hf_integration.☆56Updated 5 months ago
- Micro Llama is a small Llama based model with 300M parameters trained from scratch with $500 budget☆141Updated 10 months ago
- Training and Fine-tuning an llm in Python and PyTorch.☆41Updated last year
- Video+code lecture on building nanoGPT from scratch☆65Updated 8 months ago
- ☆74Updated last year
- RWKV in nanoGPT style☆187Updated 8 months ago
- An unofficial pytorch implementation of 'Efficient Infinite Context Transformers with Infini-attention'☆47Updated 6 months ago
- QLoRA with Enhanced Multi GPU Support☆36Updated last year
- This is the official repository for Inheritune.☆109Updated last week
- My fork os allen AI's OLMo for educational purposes.☆30Updated 2 months ago
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆115Updated last year
- Set of scripts to finetune LLMs☆36Updated 10 months ago
- RWKV, in easy to read code☆67Updated 2 months ago
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆67Updated 4 months ago
- Combining ViT and GPT-2 for image captioning. Trained on MS-COCO. The model was implemented mostly from scratch.☆38Updated last year
- A single repo with all scripts and utils to train / fine-tune the Mamba model with or without FIM☆51Updated 10 months ago
- Using open source LLMs to build synthetic datasets for direct preference optimization☆57Updated 11 months ago
- Small and Efficient Mathematical Reasoning LLMs☆71Updated last year
- This project is a collection of fine-tuning scripts to help researchers fine-tune Qwen 2 VL on HuggingFace datasets.☆63Updated 5 months ago
- GPTQLoRA: Efficient Finetuning of Quantized LLMs with GPTQ☆99Updated last year
- Load multiple LoRA modules simultaneously and automatically switch the appropriate combination of LoRA modules to generate the best answe…☆150Updated last year
- Collection of autoregressive model implementation☆81Updated last week
- Low-Rank adapter extraction for fine-tuned transformers models☆169Updated 9 months ago
- Data preparation code for Amber 7B LLM☆85Updated 9 months ago
- Pre-training code for CrystalCoder 7B LLM☆55Updated 9 months ago
- Fine-tune ModernBERT on a large Dataset with Custom Tokenizer Training☆59Updated 2 weeks ago
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆230Updated 3 months ago
- Data preparation code for CrystalCoder 7B LLM☆44Updated 9 months ago