Lightning-Universe / lightning-HivemindLinks
Lightning Training strategy for HiveMind
☆18Updated 3 weeks ago
Alternatives and similar repositories for lightning-Hivemind
Users that are interested in lightning-Hivemind are comparing it to the libraries listed below
Sorting:
- Demo of the unit_scaling library, showing how a model can be easily adapted to train in FP8.☆44Updated 10 months ago
- ☆44Updated last year
- Experiment of using Tangent to autodiff triton☆79Updated last year
- Triton Implementation of HyperAttention Algorithm☆48Updated last year
- A bunch of kernels that might make stuff slower 😉☆48Updated this week
- The source code of our work "Prepacking: A Simple Method for Fast Prefilling and Increased Throughput in Large Language Models" [AISTATS …☆59Updated 7 months ago
- ☆71Updated 2 months ago
- Write a fast kernel and run it on Discord. See how you compare against the best!☆44Updated this week
- A block oriented training approach for inference time optimization.☆33Updated 9 months ago
- A place to store reusable transformer components of my own creation or found on the interwebs☆56Updated 3 weeks ago
- A library for unit scaling in PyTorch☆125Updated 6 months ago
- Flash-Muon: An Efficient Implementation of Muon Optimizer☆121Updated last week
- Repository for Sparse Finetuning of LLMs via modified version of the MosaicML llmfoundry☆42Updated last year
- This repository contains the experimental PyTorch native float8 training UX☆223Updated 10 months ago
- ☆49Updated last year
- An extention to the GaLore paper, to perform Natural Gradient Descent in low rank subspace☆17Updated 7 months ago
- ☆78Updated 11 months ago
- The simplest implementation of recent Sparse Attention patterns for efficient LLM inference.☆62Updated 4 months ago
- ☆46Updated last week
- A safetensors extension to efficiently store sparse quantized tensors on disk☆117Updated this week
- train with kittens!☆57Updated 7 months ago
- Example ML projects that use the Determined library.☆32Updated 8 months ago
- Hacks for PyTorch☆19Updated 2 years ago
- Load compute kernels from the Hub☆144Updated this week
- A collection of reproducible inference engine benchmarks☆31Updated last month
- ☆28Updated 4 months ago
- ☆105Updated 9 months ago
- research impl of Native Sparse Attention (2502.11089)☆54Updated 3 months ago
- QuIP quantization☆52Updated last year
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆82Updated last year