校招、秋招、春招、实习好项目,带你从零动手实现支持LLama2/3和Qwen2.5的大模型推理框架。
☆498Oct 28, 2025Updated 4 months ago
Alternatives and similar repositories for KuiperLLama
Users that are interested in KuiperLLama are comparing it to the libraries listed below
Sorting:
- 校招、秋招、春招、实习好项目!带你从零实现一个高性能的深度学习推理库,支持大模型 llama2 、Unet、Yolov5、Resnet等模型的推理。Implement a high-performance deep learning inference library st…☆3,331Jun 22, 2025Updated 8 months ago
- A light llama-like llm inference framework based on the triton kernel.☆172Jan 5, 2026Updated last month
- ☆40May 11, 2025Updated 9 months ago
- ☆315Oct 9, 2024Updated last year
- 使用 CUDA C++ 实现的 llama 模型推理框架☆64Nov 8, 2024Updated last year
- 📚LeetCUDA: Modern CUDA Learn Notes with PyTorch for Beginners🐑, 200+ CUDA Kernels, Tensor Cores, HGEMM, FA-2 MMA.🎉☆9,755Updated this week
- how to optimize some algorithm in cuda.☆2,825Feb 15, 2026Updated 2 weeks ago
- CUDA 算子手撕与面试指南☆840Aug 23, 2025Updated 6 months ago
- A CUDA tutorial to make people learn CUDA program from 0☆267Jul 9, 2024Updated last year
- learning how CUDA works☆377Mar 3, 2025Updated last year
- flash attention tutorial written in python, triton, cuda, cutlass☆488Jan 20, 2026Updated last month
- llm theoretical performance analysis tools and support params, flops, memory and latency analysis.☆115Jul 11, 2025Updated 7 months ago
- ☆20Dec 29, 2023Updated 2 years ago
- 一款简单易用和高性能的AI部署框架 | An Easy-to-Use and High-Performance AI Deployment Framework☆1,743Feb 23, 2026Updated last week
- 通过onnxruntime实现yolov8在CPU和GPU上面部署☆27Aug 17, 2024Updated last year
- This is a series of GPU optimization topics. Here we will introduce how to optimize the CUDA kernel in detail. I will introduce several…☆1,244Jul 29, 2023Updated 2 years ago
- ☢️ TensorRT 2023复赛——基于TensorRT-LLM的Llama模型推断加速优化☆51Oct 20, 2023Updated 2 years ago
- how to learn PyTorch and OneFlow☆485Mar 22, 2024Updated last year
- 分层解耦的深度学习推理引擎☆79Feb 17, 2025Updated last year
- 自制基于C++的深度学习前向推理框架☆21Jun 4, 2023Updated 2 years ago
- Inference deployment of the llama3☆11Apr 21, 2024Updated last year
- ☆10Jul 18, 2024Updated last year
- Material for gpu-mode lectures☆5,773Feb 1, 2026Updated last month
- 使用 Rust 语言重新实现 https://github.com/zjhellofss/KuiperInfer 和 https://github.com/zjhellofss/kuiperdatawhale 中的深度学习推理框架。☆17Apr 9, 2024Updated last year
- A self-learning tutorail for CUDA High Performance Programing.☆900Jan 14, 2026Updated last month
- Flash Attention in ~100 lines of CUDA (forward pass only)☆1,079Dec 30, 2024Updated last year
- Llama3 Streaming Chat Sample☆22Apr 24, 2024Updated last year
- 使用 cutlass 仓库在 ada 架构上实现 fp8 的 flash attention☆79Aug 12, 2024Updated last year
- FlagGems is an operator library for large language models implemented in the Triton Language.☆904Updated this week
- ☆15Jun 22, 2025Updated 8 months ago
- 📚A curated list of Awesome LLM/VLM Inference Papers with Codes: Flash-Attention, Paged-Attention, WINT8/4, Parallelism, etc.🎉☆5,022Updated this week
- LLM notes, including model inference, transformer model structure, and llm framework code analysis notes.☆862Dec 10, 2025Updated 2 months ago
- some hpc project for learning☆26Aug 28, 2024Updated last year
- CPU Memory Compiler and Parallel programing☆26Nov 18, 2024Updated last year
- A simple high performance CUDA GEMM implementation.☆426Jan 4, 2024Updated 2 years ago
- Using pattern matcher in onnx model to match and replace subgraphs.☆81Feb 7, 2024Updated 2 years ago
- qwen2 and llama3 cpp implementation☆49Jun 7, 2024Updated last year
- ☆2,698Jan 16, 2024Updated 2 years ago
- 🤖FFPA: Extend FlashAttention-2 with Split-D, ~O(1) SRAM complexity for large headdim, 1.8x~3x↑🎉 vs SDPA EA.☆251Feb 13, 2026Updated 2 weeks ago