MooreThreads / vllm_musa
A high-throughput and memory-efficient inference and serving engine for LLMs
☆37Updated 3 months ago
Alternatives and similar repositories for vllm_musa:
Users that are interested in vllm_musa are comparing it to the libraries listed below
- DashInfer is a native LLM inference engine aiming to deliver industry-leading performance atop various hardware architectures, including …☆220Updated this week
- run ChatGLM2-6B in BM1684X☆49Updated 10 months ago
- ☆140Updated 9 months ago
- 支持中文场景的的小语言模型 llama2.c-zh☆145Updated 11 months ago
- 基于MNN-llm的安卓手机部署大语言模型:Qwen1.5-0.5B-Chat☆63Updated 9 months ago
- Run generative AI models in sophgo BM1684X☆155Updated this week
- ☆127Updated last month
- Triton Documentation in Chinese Simplified / Triton 中文文档☆52Updated 2 weeks ago
- ☆36Updated 2 months ago
- export llama to onnx☆112Updated last month
- ☆311Updated last week
- Compare different hardware platforms via the Roofline Model for LLM inference tasks.☆91Updated 10 months ago
- ☆151Updated last month
- llm-export can export llm model to onnx.☆257Updated last week
- 大模型部署实战:TensorRT-LLM, Triton Inference Server, vLLM☆26Updated 11 months ago
- run chatglm3-6b in BM1684X☆37Updated 10 months ago
- Ascend PyTorch adapter (torch_npu). Mirror of https://gitee.com/ascend/pytorch☆294Updated this week
- PaddlePaddle custom device implementaion. (『飞桨』自定义硬件接入实现)☆78Updated last week
- vLLM Documentation in Chinese Simplified / vLLM 中文文档☆25Updated 3 weeks ago
- llm theoretical performance analysis tools and support params, flops, memory and latency analysis.☆76Updated 3 weeks ago
- 使用 CUDA C++ 实现的 llama 模型推理框架☆44Updated 2 months ago
- ☢️ TensorRT 2023复赛——基于TensorRT-LLM的Llama模型推断加速优化☆45Updated last year
- ☆307Updated last month
- LLM 推理服务性能测试☆31Updated last year
- A demo built on Megrez-3B-Instruct, integrating a web search tool to enhance the model's question-and-answer capabilities.☆35Updated last month
- A MoE impl for PyTorch, [ATC'23] SmartMoE☆61Updated last year
- A high-throughput and memory-efficient inference and serving engine for LLMs☆127Updated last month
- Transformer related optimization, including BERT, GPT☆39Updated last year
- ☆57Updated 2 months ago
- ☆76Updated last year