coaxsoft / pytorch_bert
Tutorial for how to build BERT from scratch
☆91Updated 10 months ago
Alternatives and similar repositories for pytorch_bert:
Users that are interested in pytorch_bert are comparing it to the libraries listed below
- Collection of links, tutorials and best practices of how to collect the data and build end-to-end RLHF system to finetune Generative AI m…☆218Updated last year
- Well documented, unit tested, type checked and formatted implementation of a vanilla transformer - for educational purposes.☆241Updated 11 months ago
- A minimum example of aligning language models with RLHF similar to ChatGPT☆216Updated last year
- 🧠 A study guide to learn about Transformers☆11Updated last year
- This repository contains an implementation of the LLaMA 2 (Large Language Model Meta AI) model, a Generative Pretrained Transformer (GPT)…☆63Updated last year
- LLaMA 2 implemented from scratch in PyTorch☆319Updated last year
- ☆45Updated 3 years ago
- An open collection of implementation tips, tricks and resources for training large language models☆472Updated 2 years ago
- Notes and commented code for RLHF (PPO)☆85Updated last year
- LORA: Low-Rank Adaptation of Large Language Models implemented using PyTorch☆100Updated last year
- ☆81Updated last year
- A (somewhat) minimal library for finetuning language models with PPO on human feedback.☆85Updated 2 years ago
- Pre-training code for Amber 7B LLM☆166Updated 11 months ago
- ☆84Updated 6 months ago
- LLM_library is a comprehensive repository serves as a one-stop resource hands-on code, insightful summaries.☆69Updated last year
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆73Updated 5 months ago
- A set of scripts and notebooks on LLM finetunning and dataset creation☆106Updated 6 months ago
- Scripts for fine-tuning Llama2 via SFT and DPO.☆196Updated last year
- ☆97Updated last year
- LoRA and DoRA from Scratch Implementations☆200Updated last year
- Manage scalable open LLM inference endpoints in Slurm clusters☆254Updated 9 months ago
- Notes about LLaMA 2 model☆59Updated last year
- BABILong is a benchmark for LLM evaluation using the needle-in-a-haystack approach.☆197Updated last week
- ☆68Updated 2 years ago
- Training code for Baby-Llama, our submission to the strict-small track of the BabyLM challenge.☆79Updated last year
- ☆176Updated last year
- Exploring finetuning public checkpoints on filter 8K sequences on Pile☆115Updated 2 years ago
- Prune transformer layers☆68Updated 10 months ago
- Fine tune a T5 transformer model using PyTorch & Transformers🤗☆212Updated 4 years ago
- This code repository contains the code used for my "Optimizing Memory Usage for Training LLMs and Vision Transformers in PyTorch" blog po…☆91Updated last year