PrimeIntellect-ai / primeLinks
prime is a framework for efficient, globally distributed training of AI models over the internet.
☆757Updated 2 weeks ago
Alternatives and similar repositories for prime
Users that are interested in prime are comparing it to the libraries listed below
Sorting:
- OpenDiLoCo: An Open-Source Framework for Globally Distributed Low-Communication Training☆504Updated 4 months ago
- Distributed Training Over-The-Internet☆935Updated 3 weeks ago
- Minimalistic large language model 3D-parallelism training☆1,898Updated last week
- Minimalistic 4D-parallelism distributed training framework for education purpose☆1,518Updated this week
- Efficient LLM Inference over Long Sequences☆376Updated last week
- prime-rl is a codebase for decentralized async RL training at scale☆318Updated this week
- Muon is Scalable for LLM Training☆1,059Updated 2 months ago
- VPTQ, A Flexible and Extreme low-bit quantization algorithm☆639Updated last month
- Atropos is a Language Model Reinforcement Learning Environments framework for collecting and evaluating LLM trajectories through diverse …☆447Updated this week
- A throughput-oriented high-performance serving framework for LLMs☆815Updated 3 weeks ago
- [ICLR 2025] Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling☆879Updated last month
- Official implementation of Half-Quadratic Quantization (HQQ)☆818Updated this week
- OLMoE: Open Mixture-of-Experts Language Models☆773Updated 2 months ago
- Evaluate and Enhance Your LLM Deployments for Real-World Inference Needs☆317Updated this week
- KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems☆374Updated this week
- Transformers-compatible library for applying various compression algorithms to LLMs for optimized deployment with vLLM☆1,441Updated this week
- Training Large Language Model to Reason in a Continuous Latent Space☆1,138Updated 4 months ago
- ☆536Updated 9 months ago
- Single File, Single GPU, From Scratch, Efficient, Full Parameter Tuning library for "RL for LLMs"☆464Updated last week
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, spars…☆333Updated 5 months ago
- An Open Source Toolkit For LLM Distillation☆618Updated this week
- Recipes to scale inference-time compute of open models☆1,090Updated 2 weeks ago
- PyTorch per step fault tolerance (actively under development)☆302Updated this week
- LLM KV cache compression made easy☆497Updated this week
- NanoGPT (124M) in 3 minutes☆2,610Updated last week
- FlashInfer: Kernel Library for LLM Serving☆3,088Updated this week
- Advanced Quantization Algorithm for LLMs and VLMs, with support for CPU, Intel GPU, CUDA and HPU. Seamlessly integrated with Torchao, Tra…☆490Updated this week
- 🐳 Efficient Triton implementations for "Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention"☆689Updated 2 months ago
- Implementing DeepSeek R1's GRPO algorithm from scratch☆1,390Updated last month
- The Multilayer Perceptron Language Model☆549Updated 9 months ago