Apiquet / DeepLearningFrameworkFromScratchCpp
Deep Learning framework implementation with MSE, ReLU, softmax, linear layer, a feature/label generator and a mini-batch training. The main goal of this repository is to show how to develop a project in C++ by using key concepts of C++: abstract class/interface and inheritance, memory management, smart-pointers, iterator, const expression, etc.
☆21Updated last year
Alternatives and similar repositories for DeepLearningFrameworkFromScratchCpp:
Users that are interested in DeepLearningFrameworkFromScratchCpp are comparing it to the libraries listed below
- A tiny deep learning training framework implemented from scratch in C++ that follows PyTorch's API.☆48Updated 3 weeks ago
- CUDA 6大并行计算模式 代码与笔记☆60Updated 4 years ago
- Examples and exercises from the book Programming Massively Parallel Processors - A Hands-on Approach. David B. Kirk and Wen-mei W. Hwu (T…☆67Updated 4 years ago
- Implement Neural Networks in Cuda from Scratch☆22Updated 11 months ago
- Several optimization methods of half-precision general matrix vector multiplication (HGEMV) using CUDA core.☆61Updated 7 months ago
- ☆20Updated 3 years ago
- 大规模并行处理器编程实战 第二版答案☆32Updated 2 years ago
- Flash Attention in raw Cuda C beating PyTorch☆20Updated 11 months ago
- Implementation of FlashAttention in PyTorch☆143Updated 3 months ago
- Tutorials for writing high-performance GPU operators in AI frameworks.☆130Updated last year
- ☆36Updated 6 months ago
- 基于Eigen运算库的深度学习框架(支持CUDA加速)☆17Updated 3 years ago
- A light llama-like llm inference framework based on the triton kernel.☆108Updated last week
- Multiple GEMM operators are constructed with cutlass to support LLM inference.☆17Updated 6 months ago
- ☆17Updated last year
- Examples of CUDA implementations by Cutlass CuTe☆159Updated 2 months ago
- A simple deep learning framework that supports automatic differentiation and GPU acceleration.☆58Updated last year
- 使用 cutlass 仓库在 ada 架构上实现 fp8 的 flash attention☆63Updated 8 months ago
- ☆20Updated 4 years ago
- 使用 CUDA C++ 实现的 llama 模型推理框架☆50Updated 5 months ago
- CUDA 8-bit Tensor Core Matrix Multiplication based on m16n16k16 WMMA API☆30Updated last year
- ☆41Updated 3 years ago
- Optimize softmax in triton in many cases☆20Updated 7 months ago
- study of cutlass☆21Updated 5 months ago
- Standalone Flash Attention v2 kernel without libtorch dependency☆108Updated 7 months ago
- CUDA Matrix Multiplication Optimization☆181Updated 9 months ago
- ⚡️Write HGEMM from scratch using Tensor Cores with WMMA, MMA and CuTe API, Achieve Peak⚡️ Performance.☆73Updated 3 weeks ago
- Implement Flash Attention using Cute.☆76Updated 4 months ago
- 使用 cutlass 实现 flash-attention 精简版,具有教学意义☆39Updated 8 months ago
- ☆109Updated last year