ppppppppig / lite_langLinks
一个轻量化的大模型推理框架
☆18Updated last week
Alternatives and similar repositories for lite_lang
Users that are interested in lite_lang are comparing it to the libraries listed below
Sorting:
- 使用 CUDA C++ 实现的 llama 模型推理框架☆57Updated 6 months ago
- Flash Attention in ~100 lines of CUDA (forward pass only)☆10Updated 11 months ago
- TensorRT-in-Action 是一个 GitHub 代码库,提供了使用 TensorRT 的代码示例,并有对应 Jupyter Notebook。☆16Updated 2 years ago
- create your own llm inference server from scratch☆11Updated 6 months ago
- Awesome code, projects, books, etc. related to CUDA☆17Updated last month
- 使用 cutlass 仓库在 ada 架构上实现 fp8 的 flash attention☆68Updated 9 months ago
- A light llama-like llm inference framework based on the triton kernel.☆122Updated this week
- This is a repository to practice multi-thread programming in C++☆24Updated last year
- TensorRT encapsulation, learn, rewrite, practice.☆28Updated 2 years ago
- ☆36Updated 7 months ago
- ☆29Updated 6 months ago
- Optimize softmax in triton in many cases☆20Updated 9 months ago
- 使用 cutlass 实现 flash-attention 精简版,具有教学意义☆41Updated 9 months ago
- ☆58Updated 6 months ago
- ☆14Updated 9 months ago
- ☆11Updated 3 months ago
- 搜藏的希望的代码片段☆13Updated 2 years ago
- llm theoretical performance analysis tools and support params, flops, memory and latency analysis.☆92Updated last week
- A tutorial for CUDA&PyTorch☆142Updated 4 months ago
- A simple neural network inference framework☆25Updated last year
- CUDA 6大并行计算模式 代码与笔记☆61Updated 4 years ago
- ☆24Updated last year
- Multiple GEMM operators are constructed with cutlass to support LLM inference.☆18Updated 8 months ago
- ☆33Updated last year
- SGEMM optimization with cuda step by step☆19Updated last year
- Implement Flash Attention using Cute.☆85Updated 5 months ago
- Quick and Self-Contained TensorRT Custom Plugin Implementation and Integration☆60Updated last week
- ☆16Updated last year
- An onnx-based quantitation tool.☆71Updated last year
- ☢️ TensorRT 2023复赛——基于TensorRT-LLM的Llama模型推断加速优化☆48Updated last year