bashnick / transformer
A codebase implementing a simple GPT-like model from scratch based on the Attention is All You Need paper.
☆69Updated last year
Alternatives and similar repositories for transformer:
Users that are interested in transformer are comparing it to the libraries listed below
- a simplified version of Meta's Llama 3 model to be used for learning☆41Updated 11 months ago
- Tools and scripts for experimenting with Transformers: Bert, T5...☆56Updated last year
- Tutorial for how to build BERT from scratch☆93Updated 11 months ago
- Here is a Google Colab Notebook for fine-tuning Alpaca Lora (within 3 hours with a 40GB A100 GPU)☆38Updated 2 years ago
- Training and Fine-tuning an llm in Python and PyTorch.☆41Updated last year
- Implementation of Generative Pretrained Transformer Model in Tensorflow / Keras☆34Updated 11 months ago
- Fine-tune and quantize Llama-2-like models to generate Python code using QLoRA, Axolot,..☆64Updated last year
- Code implementation from my blog post: https://fkodom.substack.com/p/transformers-from-scratch-in-pytorch☆94Updated last year
- ☆75Updated last month
- Fine-tune SantaCoder for Code/Text Generation.☆192Updated 2 years ago
- nanogpt turned into a chat model☆68Updated last year
- Testing KAN-based text generation GPT models☆16Updated last year
- This is the code that went into our practical dive using mamba as information extraction☆54Updated last year
- Deep learning library implemented from scratch in numpy. Mixtral, Mamba, LLaMA, GPT, ResNet, and other experiments.☆51Updated last year
- Inference Llama 2 in one file of pure C++☆83Updated last year
- finetuning shakespeare on karpathy/nanoGPT☆19Updated 2 years ago
- LLaMA 3 is one of the most promising open-source model after Mistral, we will recreate it's architecture in a simpler manner.☆161Updated 8 months ago
- Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first app…☆168Updated last year
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆76Updated 6 months ago
- A Streamlit app running GPT-2 language model for text classification, built with Pytorch, Transformers and AWS SageMaker.☆39Updated 3 years ago
- Document Q&A on Wikipedia articles using LLMs☆76Updated last year
- [ICML 2023] "Outline, Then Details: Syntactically Guided Coarse-To-Fine Code Generation", Wenqing Zheng, S P Sharan, Ajay Kumar Jaiswal, …☆40Updated last year
- Implements a LLM similar to Meta's Llama 2 from the ground up in PyTorch, for educational purposes.☆34Updated 3 months ago
- Instruct-tune Open LLaMA / RedPajama / StableLM models on consumer hardware using QLoRA☆81Updated last year
- Learning to Program with Natural Language☆6Updated last year
- ☆73Updated 7 months ago
- ☆34Updated 2 years ago
- A simple implementation of Llama 1, 2. Llama Architecture built from scratch using PyTorch all the models are built from scratch that inc…☆13Updated last year
- Finetuning BLOOM on a single GPU using gradient-accumulation☆31Updated 2 years ago
- GPT2 finetuning with transformers 🤗☆28Updated 4 years ago