satabios / sconceLinks
E2E AutoML Model Compression Package
☆46Updated 2 months ago
Alternatives and similar repositories for sconce
Users that are interested in sconce are comparing it to the libraries listed below
Sorting:
- Cray-LM unified training and inference stack.☆22Updated 4 months ago
- Write a fast kernel and run it on Discord. See how you compare against the best!☆44Updated this week
- Collection of autoregressive model implementation☆85Updated last month
- PTX-Tutorial Written Purely By AIs (Deep Research of Openai and Claude 3.7)☆67Updated 2 months ago
- ☆48Updated this week
- VIT inference in triton because, why not?☆28Updated last year
- NanoGPT-speedrunning for the poor T4 enjoyers☆66Updated last month
- Repository for Sparse Finetuning of LLMs via modified version of the MosaicML llmfoundry☆42Updated last year
- ☆38Updated 10 months ago
- Train a SmolLM-style llm on fineweb-edu in JAX/Flax with an assortment of optimizers.☆17Updated 2 months ago
- Fork of Flame repo for training of some new stuff in development☆13Updated this week
- Make triton easier☆47Updated 11 months ago
- Lego for GRPO☆28Updated last week
- [WIP] Transformer to embed Danbooru labelsets☆13Updated last year
- Samples of good AI generated CUDA kernels☆73Updated last week
- ☆27Updated 11 months ago
- Gpu benchmark☆64Updated 4 months ago
- Cerule - A Tiny Mighty Vision Model☆67Updated 9 months ago
- Experiment of using Tangent to autodiff triton☆79Updated last year
- making the official triton tutorials actually comprehensible☆36Updated 2 months ago
- rl from zero pretrain, can it be done? we'll see.☆37Updated this week
- Official repository of Sparse ISO-FLOP Transformations for Maximizing Training Efficiency☆25Updated 10 months ago
- Demo of the unit_scaling library, showing how a model can be easily adapted to train in FP8.☆45Updated 10 months ago
- ☆44Updated last year
- Work in progress.☆68Updated last week
- ShiftAddLLM: Accelerating Pretrained LLMs via Post-Training Multiplication-Less Reparameterization☆108Updated 7 months ago
- BH hackathon☆14Updated last year
- Mixed precision training from scratch with Tensors and CUDA☆23Updated last year
- NanoGPT (124M) quality in 2.67B tokens☆28Updated last month
- ☆78Updated 11 months ago