DeepLink-org / DeepLinkExtLinks
☆13Updated 8 months ago
Alternatives and similar repositories for DeepLinkExt
Users that are interested in DeepLinkExt are comparing it to the libraries listed below
Sorting:
- ☆76Updated last year
- ☆73Updated last year
- FlagScale is a large model toolkit based on open-sourced projects.☆474Updated this week
- SGLang kernel library for NPU☆96Updated last week
- RTP-LLM: Alibaba's high-performance LLM inference engine for diverse applications.☆1,041Updated this week
- 分享AI Infra知识&代码练习:PyTorch/vLLM/SGLang框架入门⚡️、性能加速🚀、大模型基础🧠、AI软硬件🔧等☆215Updated last week
- LLM training technologies developed by kwai☆70Updated 2 weeks ago
- FlagGems is an operator library for large language models implemented in the Triton Language.☆893Updated this week
- A benchmark suited especially for deep learning operators☆42Updated 2 years ago
- [DAC2024] A Holistic Functionalization Approach to Optimizing Imperative Tensor Programs in Deep Learning☆15Updated 2 years ago
- Ascend TileLang adapter☆206Updated this week
- ☆523Updated 2 weeks ago
- Disaggregated serving system for Large Language Models (LLMs).☆771Updated 10 months ago
- Train speculative decoding models effortlessly and port them smoothly to SGLang serving.☆676Updated this week
- High Performance LLM Inference Operator Library☆695Updated this week
- A highly optimized LLM inference acceleration engine for Llama and its variants.☆906Updated this week
- GLake: optimizing GPU memory management and IO transmission.☆497Updated 10 months ago
- Materials for learning SGLang☆738Updated last month
- InternEvo is an open-sourced lightweight training framework aims to support model pre-training without the need for extensive dependencie…☆417Updated 5 months ago
- PyTorch distributed training acceleration framework☆55Updated 5 months ago
- This repository organizes materials, recordings, and schedules related to AI-infra learning meetings.☆317Updated this week
- Summary of the Specs of Commonly Used GPUs for Training and Inference of LLM☆73Updated 5 months ago
- ☆155Updated 11 months ago
- 注释的nano_vllm仓库,并且完成了MiniCPM4的适配以及注册新模型的功能☆155Updated 5 months ago
- FlagPerf is an open-source software platform for benchmarking AI chips.☆360Updated 2 months ago
- Puzzles for learning Triton, play it with minimal environment configuration!☆613Updated last month
- Ongoing research training transformer models at scale☆19Updated this week
- learning how CUDA works☆373Updated 11 months ago
- The official repo of Pai-Megatron-Patch for LLM & VLM large scale training developed by Alibaba Cloud.☆1,524Updated last month
- FlagCX is a scalable and adaptive cross-chip communication library.☆172Updated this week