yk / litter
β71Updated last year
Alternatives and similar repositories for litter:
Users that are interested in litter are comparing it to the libraries listed below
- Large scale 4D parallelism pre-training for π€ transformers in Mixture of Experts *(still work in progress)*β81Updated last year
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT trainingβ121Updated 9 months ago
- An easy-to-understand framework for LLM samplers that rewind and revise generated tokensβ118Updated 2 months ago
- Explorations into the proposal from the paper "Grokfast, Accelerated Grokking by Amplifying Slow Gradients"β95Updated 3 weeks ago
- Implementation of the Llama architecture with RLHF + Q-learningβ157Updated last year
- β56Updated last week
- Video+code lecture on building nanoGPT from scratchβ65Updated 7 months ago
- β60Updated last year
- β49Updated 10 months ago
- Ο-GPT: A New Approach to Autoregressive Modelsβ61Updated 5 months ago
- β75Updated 6 months ago
- Jax like function transformation engine but micro, microjaxβ30Updated 2 months ago
- Collection of autoregressive model implementationβ76Updated last week
- JAX Implementation of Black Forest Labs' Flux.1 family of modelsβ26Updated 2 months ago
- MiniHF is an inference, human preference data collection, and fine-tuning tool for local language models. It is intended to help the userβ¦β162Updated last week
- β62Updated 3 months ago
- β53Updated last year
- Focused on fast experimentation and simplicityβ64Updated 3 weeks ago
- β22Updated last year
- β48Updated last year
- Scripts to prep PC for development use after OS installsβ37Updated this week
- β91Updated last year
- β43Updated 2 months ago
- Cerule - A Tiny Mighty Vision Modelβ67Updated 4 months ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.β90Updated last month
- β40Updated 8 months ago
- β78Updated 9 months ago
- Full finetuning of large language models without large memory requirementsβ93Updated last year
- β29Updated last year
- β46Updated last month