zipnn / zipnnLinks
A Lossless Compression Library for AI pipelines
☆302Updated 7 months ago
Alternatives and similar repositories for zipnn
Users that are interested in zipnn are comparing it to the libraries listed below
Sorting:
- ☆280Updated this week
- Google TPU optimizations for transformers models☆134Updated 2 weeks ago
- Simple high-throughput inference library☆155Updated 8 months ago
- Load compute kernels from the Hub☆397Updated this week
- Scalable and Performant Data Loading☆364Updated this week
- A high-throughput and memory-efficient inference and serving engine for LLMs☆267Updated 2 months ago
- Simple and efficient DeepSeek V3 SFT using pipeline parallel and expert parallel, with both FP8 and BF16 trainings☆114Updated 6 months ago
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆201Updated last year
- ☆115Updated 5 months ago
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"☆155Updated last year
- LM engine is a library for pretraining/finetuning LLMs☆113Updated this week
- Accelerating your LLM training to full speed! Made with ❤️ by ServiceNow Research☆287Updated this week
- A unified library for building, evaluating, and storing speculative decoding algorithms for LLM inference in vLLM☆220Updated this week
- PyTorch implementation of models from the Zamba2 series.☆186Updated last year
- A safetensors extension to efficiently store sparse quantized tensors on disk☆238Updated this week
- Repo for "LoLCATs: On Low-Rank Linearizing of Large Language Models"☆252Updated last year
- ☆466Updated 2 months ago
- 👷 Build compute kernels☆215Updated last week
- Fault tolerance for PyTorch (HSDP, LocalSGD, DiLoCo, Streaming DiLoCo)☆475Updated this week
- Training-free Post-training Efficient Sub-quadratic Complexity Attention. Implemented with OpenAI Triton.☆148Updated 3 months ago
- ArcticInference: vLLM plugin for high-throughput, low-latency inference☆384Updated last week
- Storing long contexts in tiny caches with self-study☆236Updated 2 months ago
- ☆16Updated 2 months ago
- Manage ML configuration with pydantic☆16Updated 2 weeks ago
- Inference server benchmarking tool☆142Updated 4 months ago
- Where GPUs get cooked 👩🍳🔥☆363Updated 2 weeks ago
- Tree Attention: Topology-aware Decoding for Long-Context Attention on GPU clusters☆131Updated last year
- Formatron empowers everyone to control the format of language models' output with minimal overhead.☆234Updated 8 months ago
- ☆90Updated 7 months ago
- vLLM adapter for a TGIS-compatible gRPC server.☆50Updated this week