gmontamat / poor-mans-transformers
Implement Transformers (and Deep Learning) from scratch in NumPy
☆27Updated last year
Alternatives and similar repositories for poor-mans-transformers:
Users that are interested in poor-mans-transformers are comparing it to the libraries listed below
- A numpy implementation of the Transformer model in "Attention is All You Need"☆55Updated 9 months ago
- Well documented, unit tested, type checked and formatted implementation of a vanilla transformer - for educational purposes.☆244Updated last year
- Deep learning library implemented from scratch in numpy. Mixtral, Mamba, LLaMA, GPT, ResNet, and other experiments.☆51Updated last year
- Code Transformer neural network components piece by piece☆343Updated 2 years ago
- I will build Transformer from scratch☆68Updated 11 months ago
- LLaMA 2 implemented from scratch in PyTorch☆323Updated last year
- Tutorial for how to build BERT from scratch☆93Updated 11 months ago
- A codebase implementing a simple GPT-like model from scratch based on the Attention is All You Need paper.☆69Updated last year
- Reference implementation of Mistral AI 7B v0.1 model.☆28Updated last year
- LoRA and DoRA from Scratch Implementations☆202Updated last year
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆76Updated 6 months ago
- Use QLoRA to tune LLM in PyTorch-Lightning w/ Huggingface + MLflow☆61Updated last year
- Notes about "Attention is all you need" video (https://www.youtube.com/watch?v=bCz4OMemCcA)☆273Updated last year
- Code implementation from my blog post: https://fkodom.substack.com/p/transformers-from-scratch-in-pytorch☆94Updated last year
- Collection of autoregressive model implementation☆85Updated 2 weeks ago
- RAGs: Simple implementations of Retrieval Augmented Generation (RAG) Systems☆104Updated 3 months ago
- Fine-tune ModernBERT on a large Dataset with Custom Tokenizer Training☆66Updated 3 months ago
- 🧠 A study guide to learn about Transformers☆11Updated last year
- An extension of the nanoGPT repository for training small MOE models.☆140Updated 2 months ago
- A simplified LLAMA implementation for training and inference tasks.☆30Updated 5 months ago
- Starter pack for NeurIPS LLM Efficiency Challenge 2023.☆124Updated last year
- Contains hands-on example code for [O'reilly book "Deep Learning At Scale"](https://www.oreilly.com/library/view/deep-learning-at/9781098…☆26Updated 10 months ago
- A set of scripts and notebooks on LLM finetunning and dataset creation☆108Updated 7 months ago
- Accelerate Model Training with PyTorch 2.X, published by Packt☆43Updated 10 months ago
- The repo associated with the Manning Publication☆77Updated last month
- Building a 2.3M-parameter LLM from scratch with LLaMA 1 architecture.☆159Updated 11 months ago
- Micro Llama is a small Llama based model with 300M parameters trained from scratch with $500 budget☆150Updated last year
- Training Script and Dataset for Wealth Alpaca-LoRa☆15Updated 2 years ago
- ☆122Updated last year
- ☆47Updated last year