radixark / milesLinks
☆654Updated this week
Alternatives and similar repositories for miles
Users that are interested in miles are comparing it to the libraries listed below
Sorting:
- Checkpoint-engine is a simple middleware to update model weights in LLM inference engines☆880Updated last week
- Accelerating MoE with IO and Tile-aware Optimizations☆469Updated last week
- ByteCheckpoint: An Unified Checkpointing Library for LFMs☆256Updated 3 weeks ago
- Implementation for FP8/INT8 Rollout for RL training without performence drop.☆282Updated last month
- ☆944Updated last month
- A scalable asynchronous reinforcement learning implementation with in-flight weight updates.☆340Updated last week
- HuggingFace conversion and training library for Megatron-based models☆314Updated this week
- PyTorch-native post-training at scale☆577Updated this week
- Pytorch Distributed native training library for LLMs/VLMs with OOTB Hugging Face support☆223Updated this week
- An early research stage expert-parallel load balancer for MoE models based on linear programming.☆481Updated last month
- LLM KV cache compression made easy☆734Updated 2 weeks ago
- KernelBench: Can LLMs Write GPU Kernels? - Benchmark + Toolkit with Torch -> CUDA (+ more DSLs)☆728Updated this week
- ArcticTraining is a framework designed to simplify and accelerate the post-training process for large language models (LLMs)☆263Updated last week
- [ICLR2025 Spotlight] MagicPIG: LSH Sampling for Efficient LLM Generation☆245Updated last year
- JAX backend for SGL☆205Updated this week
- Train speculative decoding models effortlessly and port them smoothly to SGLang serving.☆597Updated this week
- [NeurIPS 2025] Simple extension on vLLM to help you speed up reasoning model without training.☆215Updated 7 months ago
- Parallel Scaling Law for Language Model — Beyond Parameter and Inference Time Scaling☆465Updated 7 months ago
- Bridge Megatron-Core to Hugging Face/Reinforcement Learning☆174Updated last week
- Genai-bench is a powerful benchmark tool designed for comprehensive token-level performance evaluation of large language model (LLM) serv…☆250Updated 3 weeks ago
- ArcticInference: vLLM plugin for high-throughput, low-latency inference☆357Updated this week
- Efficient triton implementation of Native Sparse Attention.☆257Updated 7 months ago
- Code for paper: [ICLR2025 Oral] FlexPrefill: A Context-Aware Sparse Attention Mechanism for Efficient Long-Sequence Inference☆159Updated 2 months ago
- FlexAttention based, minimal vllm-style inference engine for fast Gemma 2 inference.☆327Updated 2 months ago
- Scalable toolkit for efficient model reinforcement☆1,193Updated this week
- Memory optimized Mixture of Experts☆72Updated 5 months ago
- kernels, of the mega variety☆634Updated 3 months ago
- Async pipelined version of Verl☆124Updated 8 months ago
- KV cache compression for high-throughput LLM inference☆148Updated 10 months ago
- Allow torch tensor memory to be released and resumed later☆195Updated last month