Oneflow-Inc / oneflowLinks
OneFlow is a deep learning framework designed to be user-friendly, scalable and efficient.
☆9,361Updated 2 weeks ago
Alternatives and similar repositories for oneflow
Users that are interested in oneflow are comparing it to the libraries listed below
Sorting:
- MindSpore is a new open source deep learning training/inference framework that could be used for mobile, edge and cloud scenarios.☆4,605Updated last year
- LightSeq: A High Performance Library for Sequence Processing and Generation☆3,290Updated 2 years ago
- Jittor is a high-performance deep learning framework based on JIT compiling and meta-operators.☆3,209Updated last month
- MegEngine 是一个快速、可拓展、易于使用且支持自动求导的深度学习框架☆4,803Updated 10 months ago
- TNN: developed by Tencent Youtu Lab and Guangying Lab, a uniform deep learning inference framework for mobile、desktop and server. TNN is …☆4,577Updated 4 months ago
- A primitive library for neural network☆1,350Updated 9 months ago
- Open deep learning compiler stack for cpu, gpu and specialized accelerators☆12,584Updated this week
- Tengine is a lite, high performance, modular inference engine for embedded device☆4,489Updated 6 months ago
- TVM Documentation in Chinese Simplified / TVM 中文文档☆2,240Updated 4 months ago
- 《Machine Learning Systems: Design and Implementation》- Chinese Version☆4,603Updated last year
- DLRover: An Automatic Distributed Deep Learning System☆1,545Updated this week
- System for AI Education Resource.☆4,115Updated 10 months ago
- BladeDISC is an end-to-end DynamIc Shape Compiler project for machine learning workloads.☆891Updated 8 months ago
- OneDiff: An out-of-the-box acceleration library for diffusion models.☆1,929Updated 4 months ago
- A high performance and generic framework for distributed DNN training☆3,699Updated last year
- MACE is a deep learning inference framework optimized for mobile heterogeneous computing platforms.☆5,024Updated last year
- MiniCPM4: Ultra-Efficient LLMs on End Devices, achieving 5+ speedup on typical end-side chips☆8,261Updated 2 weeks ago
- 百亿参数的中英文双语基座大模型☆2,427Updated 2 years ago
- a fast and user-friendly runtime for transformer inference (Bert, Albert, GPT2, Decoders, etc) on CPU and GPU.☆1,532Updated last month
- Bolt is a deep learning library with high performance and heterogeneous flexibility.☆954Updated 4 months ago
- ☆1,717Updated 2 months ago
- CV-CUDA™ is an open-source, GPU accelerated library for cloud-scale image processing and computer vision.☆2,578Updated 3 months ago
- NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source compone…☆12,099Updated 2 weeks ago
- Transformer related optimization, including BERT, GPT☆6,295Updated last year
- High-performance Inference and Deployment Toolkit for LLMs and VLMs based on PaddlePaddle☆3,471Updated this week
- ncnn is a high-performance neural network inference framework optimized for the mobile platform☆22,010Updated this week
- MNN is a blazing fast, lightweight deep learning framework, battle-tested by business-critical use cases in Alibaba. Full multimodal LLM …☆12,983Updated this week
- Bagua Speeds up PyTorch☆883Updated last year
- PatrickStar enables Larger, Faster, Greener Pretrained Models for NLP and democratizes AI for everyone.☆767Updated 2 years ago
- Several simple examples for popular neural network toolkits calling custom CUDA operators.