AXERA-TECH / pulsar2-docs
The docs repository of Pulsar2 which is AXera's SoC 2rd AI toolchain. Such as AX650A, AX650N
☆11Updated last month
Related projects ⓘ
Alternatives and complementary repositories for pulsar2-docs
- Standalone Flash Attention v2 kernel without libtorch dependency☆98Updated 2 months ago
- Explore LLM model deployment based on AXera's AI chips☆54Updated last week
- TensorRT encapsulation, learn, rewrite, practice.☆24Updated 2 years ago
- NVIDIA DLA-SW, the recipes and tools for running deep learning workloads on NVIDIA DLA cores for inference applications.☆179Updated 5 months ago
- ☆57Updated this week
- A standalone GEMM kernel for fp16 activation and quantized weight, extracted from FasterTransformer☆85Updated 8 months ago
- Samples code for world class Artificial Intelligence SoCs for computer vision applications.☆215Updated last week
- ☆32Updated last month
- Common libraries for PPL projects☆29Updated 3 weeks ago
- NART = NART is not A RunTime, a deep learning inference framework.☆38Updated last year
- A set of examples around MegEngine☆30Updated 11 months ago
- ☆136Updated this week
- OneFlow->ONNX☆42Updated last year
- ☆67Updated last year
- Quick and Self-Contained TensorRT Custom Plugin Implementation and Integration☆37Updated 4 months ago
- This is a demo how to write a high performance convolution run on apple silicon☆52Updated 2 years ago
- YOLOv5 on Orin DLA☆184Updated 8 months ago
- An easy way to run, test, benchmark and tune OpenCL kernel files☆23Updated last year
- ☆12Updated 5 months ago
- MegEngine到其他框架的转换器☆67Updated last year
- A Toolkit to Help Optimize Large Onnx Model☆146Updated 5 months ago
- ☆18Updated 3 years ago
- ☆22Updated last year
- Tencent Distribution of TVM☆15Updated last year
- Offline Quantization Tools for Deploy.☆116Updated 10 months ago
- My study note for mlsys☆14Updated last week
- A simple tool that can generate TensorRT plugin code quickly.☆221Updated last year
- play gemm with tvm☆84Updated last year
- ☆18Updated 10 months ago