cccntu / LoRAnanoGPT
The simplest, fastest repository for training/finetuning medium-sized GPTs.
☆18Updated last year
Alternatives and similar repositories for LoRAnanoGPT:
Users that are interested in LoRAnanoGPT are comparing it to the libraries listed below
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Updated last year
- ☆62Updated 3 months ago
- implementation of https://arxiv.org/pdf/2312.09299☆20Updated 6 months ago
- A public implementation of the ReLoRA pretraining method, built on Lightning-AI's Pytorch Lightning suite.☆33Updated 10 months ago
- Utilities for Training Very Large Models☆57Updated 3 months ago
- [ICML 2023] "Outline, Then Details: Syntactically Guided Coarse-To-Fine Code Generation", Wenqing Zheng, S P Sharan, Ajay Kumar Jaiswal, …☆40Updated last year
- ☆49Updated 10 months ago
- Github repo for Peifeng's internship project☆12Updated last year
- A library for simplifying fine tuning with multi gpu setups in the Huggingface ecosystem.☆16Updated 2 months ago
- ☆18Updated last month
- Training hybrid models for dummies.☆16Updated this week
- CHARacter-awaRE Diffusion: Multilingual Character-Aware Encoders for Font-Aware Diffusers That Can Actually Spell☆14Updated last year
- ☆32Updated last year
- Pixel Parsing. A reproduction of OCR-free end-to-end document understanding models with open data☆21Updated 5 months ago
- QAmeleon introduces synthetic multilingual QA data using PaLM, a 540B large language model. This dataset was generated by prompt tuning P…☆34Updated last year
- [WIP] Transformer to embed Danbooru labelsets☆13Updated 9 months ago
- ☆31Updated last year
- Official repository for the paper "Approximating Two-Layer Feedforward Networks for Efficient Transformers"☆36Updated last year
- ☆12Updated last year
- GoldFinch and other hybrid transformer components☆42Updated 5 months ago
- An open source replication of the stawberry method that leverages Monte Carlo Search with PPO and or DPO☆26Updated this week
- Zeta implementation of a reusable and plug in and play feedforward from the paper "Exponentially Faster Language Modeling"☆15Updated 2 months ago
- ☆13Updated last year
- LLMs as Collaboratively Edited Knowledge Bases☆43Updated 10 months ago
- PyTorch Implementation of the paper "MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training"☆23Updated this week
- Latent Diffusion Language Models☆68Updated last year