bashnick / transformerLinks
A codebase implementing a simple GPT-like model from scratch based on the Attention is All You Need paper.
☆71Updated 2 years ago
Alternatives and similar repositories for transformer
Users that are interested in transformer are comparing it to the libraries listed below
Sorting:
- RWKV in nanoGPT style☆197Updated last year
- nanogpt turned into a chat model☆81Updated 2 years ago
- ☆53Updated last year
- GPT-2 small trained on phi-like data☆68Updated last year
- Micro Llama is a small Llama based model with 300M parameters trained from scratch with $500 budget☆169Updated 5 months ago
- A minimum example of aligning language models with RLHF similar to ChatGPT☆225Updated 2 years ago
- Training and Fine-tuning an llm in Python and PyTorch.☆43Updated 2 years ago
- Minimal code to train a Large Language Model (LLM).☆170Updated 3 years ago
- Deep learning library implemented from scratch in numpy. Mixtral, Mamba, LLaMA, GPT, ResNet, and other experiments.☆53Updated last year
- Experimenting with small language models☆76Updated 2 years ago
- Inference Llama 2 in one file of pure C++☆87Updated 2 years ago
- A Streamlit app running GPT-2 language model for text classification, built with Pytorch, Transformers and AWS SageMaker.☆39Updated 4 years ago
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆77Updated last year
- ☆42Updated last year
- tinygrad port of the RWKV large language model.☆45Updated 10 months ago
- ☆88Updated 2 years ago
- Reweight GPT - a simple neural network using transformer architecture for next character prediction☆57Updated 2 years ago
- Like picoGPT but for BERT.☆51Updated 2 years ago
- Well documented, unit tested, type checked and formatted implementation of a vanilla transformer - for educational purposes.☆282Updated last year
- minimal scripts for 24GB VRAM GPUs. training, inference, whatever☆50Updated last month
- ☆86Updated 2 years ago
- ☆17Updated last year
- Small and Efficient Mathematical Reasoning LLMs☆73Updated 2 years ago
- Pre-training code for CrystalCoder 7B LLM☆57Updated last year
- The simplest, fastest repository for training/finetuning medium-sized GPTs with LoRA support.☆29Updated last year
- ☆42Updated 2 years ago
- Inference on CPU code for LLaMA models☆137Updated 2 years ago
- LORA: Low-Rank Adaptation of Large Language Models implemented using PyTorch☆122Updated 2 years ago
- Documented and Unit Tested educational Deep Learning framework with Autograd from scratch.☆122Updated last year
- ☆84Updated 2 years ago