luliyucoordinate / flash-attention-minimalLinks
Flash Attention in ~100 lines of CUDA (forward pass only)
☆10Updated last year
Alternatives and similar repositories for flash-attention-minimal
Users that are interested in flash-attention-minimal are comparing it to the libraries listed below
Sorting:
- Awesome code, projects, books, etc. related to CUDA☆17Updated last week
- TensorRT-in-Action 是一个 GitHub 代码库,提供了使用 TensorRT 的代码示例,并有对应 Jupyter Notebook。☆16Updated 2 years ago
- 使用 CUDA C++ 实现的 llama 模型推理框架☆57Updated 7 months ago
- A light llama-like llm inference framework based on the triton kernel.☆128Updated last week
- ☆26Updated last year
- 一个轻量化的大模型推理框架☆19Updated last month
- TensorRT encapsulation, learn, rewrite, practice.☆28Updated 2 years ago
- create your own llm inference server from scratch☆12Updated 7 months ago
- ffmpeg+cuvid+tensorrt+multicamera☆12Updated 5 months ago
- Optimize softmax in triton in many cases☆21Updated 9 months ago
- Implement Flash Attention using Cute.☆87Updated 6 months ago
- ☆16Updated last year
- ☢️ TensorRT 2023复赛——基于TensorRT-LLM的Llama模型推断加速优化☆48Updated last year
- 🎉My Collections of CUDA Kernels~☆11Updated last year
- ☆14Updated 10 months ago
- ☆34Updated last year
- ☆28Updated last month
- ☆36Updated 8 months ago
- Multiple GEMM operators are constructed with cutlass to support LLM inference.☆18Updated 9 months ago
- 搜藏的希望的代码片段☆13Updated 2 years ago
- async inference for machine learning model☆26Updated 2 years ago
- This is a repository to practice multi-thread programming in C++☆24Updated last year
- Llama3 Streaming Chat Sample☆22Updated last year
- 使用 cutlass 仓库在 ada 架构上实现 fp8 的 flash attention☆71Updated 10 months ago
- ☆30Updated 7 months ago
- HunyuanDiT with TensorRT and libtorch☆17Updated last year
- ☆135Updated last year
- NVIDIA TensorRT Hackathon 2023复赛选题:通义千问Qwen-7B用TensorRT-LLM模型搭建及优化☆42Updated last year
- ☆58Updated 7 months ago
- learn TensorRT from scratch🥰☆15Updated 8 months ago