abacaj / awesome-transformers
A curated list of awesome transformer models.
☆641Updated last year
Alternatives and similar repositories for awesome-transformers:
Users that are interested in awesome-transformers are comparing it to the libraries listed below
- ☆589Updated last year
- [NeurIPS 22] [AAAI 24] Recurrent Transformer-based long-context architecture.☆760Updated 5 months ago
- A collection of open-source dataset to train instruction-following LLMs (ChatGPT,LLaMA,Alpaca)☆1,113Updated last year
- The complete training code of the open-source high-performance Llama model, including the full process from pre-training to RLHF.☆43Updated last year
- Salesforce open-source LLMs with 8k sequence length.☆716Updated last month
- Codes for "Chameleon: Plug-and-Play Compositional Reasoning with Large Language Models".☆1,123Updated last year
- Run inference on MPT-30B using CPU☆575Updated last year
- Ask Me Anything language model prompting☆546Updated last year
- An open collection of implementation tips, tricks and resources for training large language models☆471Updated 2 years ago
- A crude RLHF layer on top of nanoGPT with Gumbel-Softmax trick☆288Updated last year
- A collection of modular datasets generated by GPT-4, General-Instruct - Roleplay-Instruct - Code-Instruct - and Toolformer☆1,627Updated last year
- Build, evaluate, understand, and fix LLM-based apps☆487Updated last year
- An open collection of methodologies to help with successful training of large language models.☆480Updated last year
- OpenICL is an open-source framework to facilitate research, development, and prototyping of in-context learning.☆552Updated last year
- Generate textbook-quality synthetic LLM pretraining data☆498Updated last year
- OpenAlpaca: A Fully Open-Source Instruction-Following Model Based On OpenLLaMA☆301Updated last year
- MultimodalC4 is a multimodal extension of c4 that interleaves millions of images with text.☆921Updated this week
- LLM papers I'm reading, mostly on inference and model compression☆715Updated last year
- LaMini-LM: A Diverse Herd of Distilled Models from Large-Scale Instructions☆820Updated last year
- Reading list of Instruction-tuning. A trend starts from Natrural-Instruction (ACL 2022), FLAN (ICLR 2022) and T0 (ICLR 2022).☆766Updated last year
- C++ implementation for BLOOM☆809Updated last year
- A central, open resource for data and tools related to chain-of-thought reasoning in large language models. Developed @ Samwald research …☆948Updated 3 months ago
- Understanding large language models☆118Updated 2 years ago
- A curated list of awesome instruction tuning datasets, models, papers and repositories.☆330Updated last year
- Fast & Simple repository for pre-training and fine-tuning T5-style models☆1,000Updated 7 months ago
- This repository contains code and tooling for the Abacus.AI LLM Context Expansion project. Also included are evaluation scripts and bench…☆584Updated last year
- [NeurIPS 2023] MeZO: Fine-Tuning Language Models with Just Forward Passes. https://arxiv.org/abs/2305.17333☆1,095Updated last year
- Human preference data for "Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback"☆1,707Updated last year
- Fine-tune mistral-7B on 3090s, a100s, h100s☆709Updated last year
- Implementation of RETRO, Deepmind's Retrieval based Attention net, in Pytorch☆859Updated last year