FlyAIBox / dcu-in-actionLinks
国产加速卡-海光DCU实战(大模型训练、微调、推理 等)
☆55Updated 3 months ago
Alternatives and similar repositories for dcu-in-action
Users that are interested in dcu-in-action are comparing it to the libraries listed below
Sorting:
- llm-inference is a platform for publishing and managing llm inference, providing a wide range of out-of-the-box features for model deploy…☆87Updated last year
- Manages vllm-nccl dependency☆17Updated last year
- CPM.cu is a lightweight, high-performance CUDA implementation for LLMs, optimized for end-device inference and featuring cutting-edge tec…☆201Updated last month
- Model compression toolkit engineered for enhanced usability, comprehensiveness, and efficiency.☆197Updated this week
- Omni_Infer is a suite of inference accelerators designed for the Ascend NPU platform, offering native support and an expanding feature se…☆84Updated this week
- vLLM Router☆50Updated last year
- Open deep learning compiler stack for cpu, gpu and specialized accelerators☆19Updated last week
- A benchmarking tool for comparing different LLM API providers' DeepSeek model deployments.☆30Updated 7 months ago
- 配合 HAI Platform 使用的集成化用户界面☆53Updated 2 years ago
- The driver for LMCache core to run in vLLM☆56Updated 9 months ago
- DeepTrace: A lightweight, scalable real-time diagnostic and analysis tool for distributed training tasks.☆17Updated last week
- ☆56Updated 11 months ago
- Efficient, Flexible, and Highly Fault-Tolerant Model Service Management Based on SGLang☆59Updated last year
- ☆25Updated 2 years ago
- ☆25Updated 10 months ago
- TensorRT LLM Benchmark Configuration☆13Updated last year
- Transformer related optimization, including BERT, GPT☆17Updated 2 years ago
- GLM Series Edge Models☆153Updated 5 months ago
- ☆79Updated last year
- DashInfer is a native LLM inference engine aiming to deliver industry-leading performance atop various hardware architectures, including …☆267Updated 3 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆16Updated last year
- Delta-CoMe can achieve near loss-less 1-bit compressin which has been accepted by NeurIPS 2024☆57Updated 11 months ago
- ☆125Updated this week
- LLM 推理服务性能测试☆44Updated last year
- patches for huggingface transformers to save memory☆31Updated 5 months ago
- Official Implementation of APB (ACL 2025 main Oral)☆31Updated 8 months ago
- ☆112Updated last year
- Compare different hardware platforms via the Roofline Model for LLM inference tasks.☆119Updated last year
- DLBlas: clean and efficient kernels☆23Updated this week
- 官方transformers源码解析。AI大模型时代,pytorch、transformer是新操作系统,其他都是运行在其上面的软件。☆17Updated 2 years ago