zhihu / ZhiLight
A highly optimized LLM inference acceleration engine for Llama and its variants.
☆884Updated this week
Alternatives and similar repositories for ZhiLight:
Users that are interested in ZhiLight are comparing it to the libraries listed below
- An acceleration library that supports arbitrary bit-width combinatorial quantization operations☆223Updated 7 months ago
- TVM Documentation in Chinese Simplified / TVM 中文文档☆1,188Updated 3 weeks ago
- RTP-LLM: Alibaba's high-performance LLM inference engine for diverse applications.☆718Updated 3 months ago
- Unified KV Cache Compression Methods for Auto-Regressive Models☆1,051Updated 4 months ago
- adds Sequence Parallelism into LLaMA-Factory☆482Updated last week
- minimal-cost for training 0.5B R1-Zero☆716Updated 2 weeks ago
- FlagPerf is an open-source software platform for benchmarking AI chips.☆331Updated 3 months ago
- ☆329Updated 3 months ago
- FlagScale is a large model toolkit based on open-sourced projects.☆270Updated this week
- xDiT: A Scalable Inference Engine for Diffusion Transformers (DiTs) with Massive Parallelism☆1,920Updated 2 weeks ago
- Higher performance OpenAI LLM service than vLLM serve: A pure C++ high-performance OpenAI LLM service implemented with GPRS+TensorRT-LLM+…☆133Updated this week
- MIXQ: Taming Dynamic Outliers in Mixed-Precision Quantization by Online Prediction☆88Updated 6 months ago
- GLake: optimizing GPU memory management and IO transmission.☆457Updated last month
- A fast communication-overlapping library for tensor/expert parallelism on GPUs.☆915Updated 3 weeks ago
- ☆490Updated this week
- optimized BERT transformer inference on NVIDIA GPU. https://arxiv.org/abs/2210.03052☆473Updated last year
- DashInfer is a native LLM inference engine aiming to deliver industry-leading performance atop various hardware architectures, including …☆245Updated this week
- Community maintained hardware plugin for vLLM on Ascend☆605Updated this week
- Disaggregated serving system for Large Language Models (LLMs).☆580Updated last month
- FlagGems is an operator library for large language models implemented in Triton Language.☆516Updated this week
- Materials for learning SGLang☆406Updated 2 weeks ago
- Deep Learning Deployment Framework: Supports tf/torch/trt/trtllm/vllm and other NN frameworks. Support dynamic batching, and streaming mo…☆156Updated last month
- DeepRetrieval - Hacking 🔥Real Search Engines and Retrievers with LLM via RL☆487Updated this week
- ☆67Updated 6 months ago
- The repository has collected a batch of noteworthy MLSys bloggers (Algorithms/Systems)☆228Updated 4 months ago
- [EMNLP 2024 Industry Track] This is the official PyTorch implementation of "LLMC: Benchmarking Large Language Model Quantization with a V…☆472Updated this week
- A self-learning tutorail for CUDA High Performance Programing.☆626Updated 3 weeks ago
- Puzzles for learning Triton, play it with minimal environment configuration!☆302Updated 5 months ago
- DLRover: An Automatic Distributed Deep Learning System☆1,435Updated this week
- High-performance inference framework for large language models, focusing on efficiency, flexibility, and availability.☆1,112Updated this week