Andrei-Aksionov / nanoGPTplusLinks
☆51Updated last year
Alternatives and similar repositories for nanoGPTplus
Users that are interested in nanoGPTplus are comparing it to the libraries listed below
Sorting:
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆78Updated last year
- Training and Fine-tuning an llm in Python and PyTorch.☆43Updated 2 years ago
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆231Updated last year
- ☆94Updated 2 years ago
- Minimal example scripts of the Hugging Face Trainer, focused on staying under 150 lines☆195Updated last year
- FineTune LLMs in few lines of code (Text2Text, Text2Speech, Speech2Text)☆242Updated last year
- ☆197Updated last year
- 🕹️ Performance Comparison of MLOps Engines, Frameworks, and Languages on Mainstream AI Models.☆139Updated last year
- Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first app…☆169Updated last year
- Minimal code to train a Large Language Model (LLM).☆172Updated 3 years ago
- inference code for mixtral-8x7b-32kseqlen☆102Updated last year
- 🦖 X—LLM: Cutting Edge & Easy LLM Finetuning☆406Updated last year
- A bagel, with everything.☆324Updated last year
- Pre-training code for Amber 7B LLM☆169Updated last year
- Some simple scripts that I use day-to-day when working with LLMs and Huggingface Hub☆160Updated 2 years ago
- A set of scripts and notebooks on LLM finetunning and dataset creation☆110Updated last year
- Reweight GPT - a simple neural network using transformer architecture for next character prediction☆58Updated 2 years ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆266Updated last year
- Supervised instruction finetuning for LLM with HF trainer and Deepspeed☆36Updated 2 years ago
- 📚 Datasets and models for instruction-tuning☆237Updated 2 years ago
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.☆82Updated 2 years ago
- ☆88Updated last year
- experiments with inference on llama☆103Updated last year
- Train your own small bitnet model☆75Updated last year
- Manage scalable open LLM inference endpoints in Slurm clusters☆273Updated last year
- The repository for the code of the UltraFastBERT paper☆518Updated last year
- Reimplementation of the task generation part from the Alpaca paper☆118Updated 2 years ago
- Micro Llama is a small Llama based model with 300M parameters trained from scratch with $500 budget☆161Updated 2 months ago
- Toolkit for attaching, training, saving and loading of new heads for transformer models☆289Updated 7 months ago
- A comprehensive deep dive into the world of tokens☆226Updated last year