bashnick / transformerLinks
A codebase implementing a simple GPT-like model from scratch based on the Attention is All You Need paper.
☆71Updated 2 years ago
Alternatives and similar repositories for transformer
Users that are interested in transformer are comparing it to the libraries listed below
Sorting:
- nanogpt turned into a chat model☆80Updated 2 years ago
- Reweight GPT - a simple neural network using transformer architecture for next character prediction☆57Updated 2 years ago
- Training and Fine-tuning an llm in Python and PyTorch.☆43Updated 2 years ago
- This is the code that went into our practical dive using mamba as information extraction☆57Updated 2 years ago
- RWKV in nanoGPT style☆197Updated last year
- Code base for internal reward models and PPO training☆24Updated 2 years ago
- ☆148Updated last year
- ☆41Updated last year
- Tools and scripts for experimenting with Transformers: Bert, T5...☆61Updated 2 years ago
- Micro Llama is a small Llama based model with 300M parameters trained from scratch with $500 budget☆169Updated 5 months ago
- ☆86Updated 2 years ago
- Create an AI capable of solving reasoning tasks it has never seen before☆96Updated last year
- Inference RWKV v7 in pure C.☆43Updated 3 months ago
- A minimum example of aligning language models with RLHF similar to ChatGPT☆225Updated 2 years ago
- LLaMA 3 is one of the most promising open-source model after Mistral, we will recreate it's architecture in a simpler manner.☆197Updated last year
- ☆65Updated 2 years ago
- GPT-2 small trained on phi-like data☆68Updated last year
- Experimenting with small language models☆76Updated 2 years ago
- A Very Simple Vector Database☆15Updated 2 years ago
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆77Updated last year
- Port of Andrej Karpathy's nanoGPT to Apple MLX framework.☆117Updated last year
- Pre-training code for CrystalCoder 7B LLM☆57Updated last year
- Evaluation of bm42 sparse indexing algorithm☆72Updated last year
- Inference of Mamba and Mamba2 models in pure C☆196Updated last week
- Minimal code to train a Large Language Model (LLM).☆170Updated 3 years ago
- ☆31Updated 2 years ago
- The simplest, fastest repository for training/finetuning medium-sized xLSTMs.☆41Updated last year
- Here is a Google Colab Notebook for fine-tuning Alpaca Lora (within 3 hours with a 40GB A100 GPU)☆38Updated 2 years ago
- Train your own small bitnet model☆77Updated last year
- A really tiny autograd engine☆99Updated 8 months ago