warner-benjamin / commented-transformers
Highly commented implementations of Transformers in PyTorch
☆132Updated last year
Alternatives and similar repositories for commented-transformers:
Users that are interested in commented-transformers are comparing it to the libraries listed below
- Minimal example scripts of the Hugging Face Trainer, focused on staying under 150 lines☆198Updated 9 months ago
- A miniture AI training framework for PyTorch☆39Updated last month
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.☆82Updated last year
- ☆92Updated last year
- Source notebook code for the course, stripped of all information. Please consider puchasing the course at https://store.walkwithfastai.co…☆36Updated last year
- A set of scripts and notebooks on LLM finetunning and dataset creation☆104Updated 5 months ago
- An introduction to LLM Sampling☆75Updated 2 months ago
- ☆165Updated 9 months ago
- ☆118Updated 4 months ago
- ☆77Updated 9 months ago
- ☆199Updated last year
- Train fastai models faster (and other useful tools)☆64Updated 8 months ago
- ML/DL Math and Method notes☆58Updated last year
- Fast, Modern, Memory Efficient, and Low Precision PyTorch Optimizers☆84Updated 7 months ago
- Helpers and such for working with Lambda Cloud☆52Updated last year
- Just some miscellaneous utility functions / decorators / modules related to Pytorch and Accelerate to help speed up implementation of new…☆120Updated 7 months ago
- A comprehensive deep dive into the world of tokens☆220Updated 8 months ago
- ☆40Updated 9 months ago
- [WIP] A 🔥 interface for running code in the cloud☆86Updated 2 years ago
- Explorations into the proposal from the paper "Grokfast, Accelerated Grokking by Amplifying Slow Gradients"☆96Updated 2 months ago
- A Jax-based library for designing and training transformer models from scratch.☆281Updated 6 months ago
- ☆149Updated 6 months ago
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆230Updated 4 months ago
- ☆301Updated 8 months ago
- ☆150Updated 3 months ago
- Outlining techniques for improving the training performance of your PyTorch model without compromising its accuracy☆127Updated last year
- The Batched API provides a flexible and efficient way to process multiple requests in a batch, with a primary focus on dynamic batching o…☆123Updated 2 months ago
- ☆75Updated 7 months ago
- Functional local implementations of main model parallelism approaches☆95Updated 2 years ago
- ☆88Updated 2 weeks ago