Blaizzy / Coding-LLMs-from-scratchLinks
☆36Updated last year
Alternatives and similar repositories for Coding-LLMs-from-scratch
Users that are interested in Coding-LLMs-from-scratch are comparing it to the libraries listed below
Sorting:
- Micro Llama is a small Llama based model with 300M parameters trained from scratch with $500 budget☆169Updated 5 months ago
- inference code for mixtral-8x7b-32kseqlen☆105Updated 2 years ago
- a simplified version of Google's Gemma model to be used for learning☆26Updated last year
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆232Updated last year
- Video+code lecture on building nanoGPT from scratch☆68Updated last year
- One click templates for inferencing Language Models☆228Updated 2 months ago
- ☆137Updated last year
- Some simple scripts that I use day-to-day when working with LLMs and Huggingface Hub☆161Updated 2 years ago
- Full finetuning of large language models without large memory requirements☆94Updated 4 months ago
- Various installation guides for Large Language Models☆77Updated 9 months ago
- ☆52Updated 2 years ago
- Low-Rank adapter extraction for fine-tuned transformers models☆180Updated last year
- ☆38Updated last year
- an implementation of Self-Extend, to expand the context window via grouped attention☆119Updated 2 years ago
- Port of Andrej Karpathy's nanoGPT to Apple MLX framework.☆118Updated last year
- Scripts to create your own moe models using mlx☆90Updated last year
- Cerule - A Tiny Mighty Vision Model☆68Updated 2 months ago
- GRDN.AI app for garden optimization☆69Updated 2 months ago
- Fast parallel LLM inference for MLX☆245Updated last year
- A simple MLX implementation for pretraining LLMs on Apple Silicon.☆85Updated 5 months ago
- A set of scripts and notebooks on LLM finetunning and dataset creation☆116Updated last year
- Let's create synthetic textbooks together :)☆76Updated 2 years ago
- Function Calling Benchmark & Testing☆92Updated last year
- nanogpt turned into a chat model☆81Updated 2 years ago
- Convenient wrapper for fine-tuning and inference of Large Language Models (LLMs) with several quantization techniques (GTPQ, bitsandbytes…☆146Updated 2 years ago
- Distributed Inference for mlx LLm☆100Updated last year
- ☆86Updated 2 years ago
- ☆127Updated 10 months ago
- Zeus LLM Trainer is a rewrite of Stanford Alpaca aiming to be the trainer for all Large Language Models☆70Updated 2 years ago
- An introduction to LLM Sampling☆79Updated last year