Oneflow-Inc / oneflow-liteLinks
☆18Updated last year
Alternatives and similar repositories for oneflow-lite
Users that are interested in oneflow-lite are comparing it to the libraries listed below
Sorting:
- OneFlow->ONNX☆43Updated 2 years ago
- ☆11Updated 3 months ago
- OneFlow Serving☆20Updated 7 months ago
- ☆23Updated 2 years ago
- A toolkit for developers to simplify the transformation of nn.Module instances. It's now corresponding to Pytorch.fx.☆13Updated 2 years ago
- CVFusion is an open-source deep learning compiler to fuse the OpenCV operators.☆31Updated 3 years ago
- study of cutlass☆22Updated last year
- An easy way to run, test, benchmark and tune OpenCL kernel files☆24Updated 2 years ago
- ☆33Updated 9 months ago
- GPTQ inference TVM kernel☆39Updated last year
- Decoding Attention is specially optimized for MHA, MQA, GQA and MLA using CUDA core for the decoding stage of LLM inference.☆45Updated 5 months ago
- ☆15Updated 3 years ago
- A Triton JIT runtime and ffi provider in C++☆29Updated last week
- Multiple GEMM operators are constructed with cutlass to support LLM inference.☆20Updated 3 months ago
- A standalone GEMM kernel for fp16 activation and quantized weight, extracted from FasterTransformer☆96Updated 2 months ago
- This is a demo how to write a high performance convolution run on apple silicon☆57Updated 3 years ago
- NVIDIA TensorRT Hackathon 2023复赛选题:通义千问Qwen-7B用TensorRT-LLM模型搭建及优化☆43Updated 2 years ago
- Standalone Flash Attention v2 kernel without libtorch dependency☆112Updated last year
- ☆38Updated last year
- ☆24Updated 2 years ago
- 分层解耦的深度学习推理引擎☆76Updated 8 months ago
- High Performance FP8 GEMM Kernels for SM89 and later GPUs.☆20Updated 9 months ago
- Yet another Polyhedra Compiler for DeepLearning☆19Updated 2 years ago
- Datasets, Transforms and Models specific to Computer Vision☆90Updated last year
- A high-throughput and memory-efficient inference and serving engine for LLMs☆16Updated last year
- Performance of the C++ interface of flash attention and flash attention v2 in large language model (LLM) inference scenarios.☆42Updated 8 months ago
- ☆12Updated 2 years ago
- Transformer related optimization, including BERT, GPT☆17Updated 2 years ago
- 使用 cutlass 仓库在 ada 架构上实现 fp8 的 flash attention☆78Updated last year
- ☆60Updated 11 months ago