A light llama-like llm inference framework based on the triton kernel.
☆180Jan 5, 2026Updated 3 months ago
Alternatives and similar repositories for lite_llama
Users that are interested in lite_llama are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- ☆46Mar 4, 2026Updated last month
- llm theoretical performance analysis tools and support params, flops, memory and latency analysis.☆117Jul 11, 2025Updated 9 months ago
- 校招、秋招、春招、实习好项目,带你从零动手实现支持LLama2/3和Qwen2.5的大模型推理框架。☆527Oct 28, 2025Updated 5 months ago
- LLM notes, including model inference, transformer model structure, and llm framework code analysis notes.☆875Updated this week
- 一个轻量化的大模型推理框架☆22May 26, 2025Updated 10 months ago
- Wordpress hosting with auto-scaling - Free Trial • AdFully Managed hosting for WordPress and WooCommerce businesses that need reliable, auto-scalable performance. Cloudways SafeUpdates now available.
- 校招、秋招、春招、实习好项目!带你从零实现一个高性能的深度学习推理库,支持大模型 llama2 、Unet、Yolov5、Resnet等模型的推理。Implement a high-performance deep learning inference library st…☆3,396Jun 22, 2025Updated 9 months ago
- 68th palce solution in Kaggle Humpback Whale Identification.☆11Jul 6, 2023Updated 2 years ago
- FlagGems is an operator library for large language models implemented in the Triton Language.☆953Updated this week
- 📚LeetCUDA: Modern CUDA Learn Notes with PyTorch for Beginners🐑, 200+ CUDA Kernels, Tensor Cores, HGEMM, FA-2 MMA.🎉☆10,290Apr 12, 2026Updated last week
- ☆136Mar 5, 2026Updated last month
- 使用 CUDA C++ 实现的 llama 模型推理框架☆65Nov 8, 2024Updated last year
- 本仓库在OpenVINO推理框架下部 署Nanodet检测算法,并重写预处理和后处理部分,具有超高性能!让你在Intel CPU平台上的检测速度起飞! 并基于NNCF和PPQ工具将模型量化(PTQ)至int8精度,推理速度更快!☆16Jun 14, 2023Updated 2 years ago
- 深度学习系统笔记,包含深度学习数学基础知识、神经网络基础部件详解、深度学习炼丹策略、模型压缩算法详解。☆516Dec 11, 2025Updated 4 months ago
- 一款简单易用和高性能的AI部署框架 | An Easy-to-Use and High-Performance AI Deployment Framework☆1,782Apr 12, 2026Updated last week
- Simple, predictable pricing with DigitalOcean hosting • AdAlways know what you'll pay with monthly caps and flat pricing. Enterprise-grade infrastructure trusted by 600k+ customers.
- A single-file educational implementation for understanding vLLM's core concepts and running LLM inference.☆44Apr 7, 2026Updated last week
- A CUDA tutorial to make people learn CUDA program from 0☆276Jul 9, 2024Updated last year
- how to learn PyTorch and OneFlow☆494Mar 22, 2024Updated 2 years ago
- LLM Inference Engine: High-performance CUDA-accelerated framework for large language model inference A cutting-edge, open-source impleme…☆11Sep 29, 2024Updated last year
- From Minimal GEMM to Everything☆199Feb 10, 2026Updated 2 months ago
- how to optimize some algorithm in cuda.☆2,925Apr 9, 2026Updated last week
- ☆15Jun 22, 2025Updated 9 months ago
- A tiny yet powerful LLM inference system tailored for researching purpose. vLLM-equivalent performance with only 2k lines of code (2% of …☆323Jun 10, 2025Updated 10 months ago
- 使用 cutlass 仓库在 ada 架构上实现 fp8 的 flash attention☆82Aug 12, 2024Updated last year
- Wordpress hosting with auto-scaling - Free Trial • AdFully Managed hosting for WordPress and WooCommerce businesses that need reliable, auto-scalable performance. Cloudways SafeUpdates now available.
- Optimize softmax in triton in many cases☆23Sep 6, 2024Updated last year
- 通过onnxruntime实现yolov8在CPU和GPU上面部署☆27Aug 17, 2024Updated last year
- Quantize yolov5 using pytorch_quantization.🚀🚀🚀☆14Oct 24, 2023Updated 2 years ago
- 🤖FFPA: Extend FlashAttention-2 with Split-D, ~O(1) SRAM complexity for large headdim, 1.8x~3x↑🎉 vs SDPA EA.☆260Feb 13, 2026Updated 2 months ago
- Build a Claude Code–like CLI coding agent from scratch.☆135Jan 22, 2026Updated 2 months ago
- flash attention tutorial written in python, triton, cuda, cutlass☆502Jan 20, 2026Updated 2 months ago
- An onnx-based quantitation tool.☆71Jan 8, 2024Updated 2 years ago
- CUDA SGEMM optimization note☆15Oct 31, 2023Updated 2 years ago
- https://github.com/shouxieai/hard_decode_trt windows编译版本☆13Sep 8, 2022Updated 3 years ago
- GPUs on demand by Runpod - Special Offer Available • AdRun AI, ML, and HPC workloads on powerful cloud GPUs—without limits or wasted spend. Deploy GPUs in under a minute and pay by the second.
- Analyze the inference of Large Language Models (LLMs). Analyze aspects like computation, storage, transmission, and hardware roofline mod…☆635Sep 11, 2024Updated last year
- 注释的nano_vllm仓库,并且完成了MiniCPM4的适配以及注册新模型的功能☆179Aug 11, 2025Updated 8 months ago
- 【A common used C++ & Python DAG framework】 一个通用的、无 三方依赖的、跨平台的、收录于awesome-cpp的、基于流图的并行计算框架。欢迎star & fork & 交流☆2,262Apr 11, 2026Updated last week
- Material for gpu-mode lectures☆5,945Feb 1, 2026Updated 2 months ago
- A simple neural network inference framework☆25Aug 1, 2023Updated 2 years ago
- ☆79May 16, 2023Updated 2 years ago
- ☆48Mar 27, 2023Updated 3 years ago