PrimeIntellect-ai / pi-quantLinks
SIMD quantization kernels
☆92Updated 2 months ago
Alternatives and similar repositories for pi-quant
Users that are interested in pi-quant are comparing it to the libraries listed below
Sorting:
- PCCL (Prime Collective Communications Library) implements fault tolerant collective communications over IP☆138Updated 2 months ago
- Simple Transformer in Jax☆139Updated last year
- Quantized LLM training in pure CUDA/C++.☆220Updated this week
- look how they massacred my boy☆63Updated last year
- Official CLI and Python SDK for Prime Intellect - access GPU compute, remote sandboxes, RL environments, and distributed training infrast…☆113Updated this week
- NSA Triton Kernels written with GPT5 and Opus 4.1☆65Updated 3 months ago
- DeMo: Decoupled Momentum Optimization☆197Updated last year
- Storing long contexts in tiny caches with self-study☆218Updated last month
- NanoGPT-speedrunning for the poor T4 enjoyers☆73Updated 7 months ago
- Plotting (entropy, varentropy) for small LMs☆99Updated 6 months ago
- Compiling useful links, papers, benchmarks, ideas, etc.☆45Updated 8 months ago
- in this repository, i'm going to implement increasingly complex llm inference optimizations☆70Updated 6 months ago
- A zero-to-one guide on scaling modern transformers with n-dimensional parallelism.☆105Updated 2 months ago
- The Automated LLM Speedrunning Benchmark measures how well LLM agents can reproduce previous innovations and discover new ones in languag…☆112Updated last month
- ☆234Updated 5 months ago
- Training-Ready RL Environments + Evals☆182Updated this week
- rl from zero pretrain, can it be done? yes.☆281Updated 2 months ago
- smolLM with Entropix sampler on pytorch☆149Updated last year
- an open source reproduction of NVIDIA's nGPT (Normalized Transformer with Representation Learning on the Hypersphere)☆108Updated 8 months ago
- peer-to-peer compute and intelligence network that enables decentralized AI development at scale☆135Updated 3 weeks ago
- A simple MLX implementation for pretraining LLMs on Apple Silicon.☆84Updated 3 months ago
- Learning about CUDA by writing PTX code.☆148Updated last year
- ☆40Updated last year
- A graph visualization of attention☆57Updated 6 months ago
- smol models are fun too☆92Updated last year
- train entropix like a champ!☆20Updated last year
- A tree-based prefix cache library that allows rapid creation of looms: hierarchal branching pathways of LLM generations.☆76Updated 9 months ago
- Modded vLLM to run pipeline parallelism over public networks☆40Updated 6 months ago
- Simple & Scalable Pretraining for Neural Architecture Research☆302Updated last month
- Ultra low overhead NVIDIA GPU telemetry plugin for telegraf with memory temperature readings.☆63Updated last year