intel / ai-reference-models
Intel® AI Reference Models: contains Intel optimizations for running deep learning workloads on Intel® Xeon® Scalable processors and Intel® Data Center GPUs
☆706Updated this week
Alternatives and similar repositories for ai-reference-models:
Users that are interested in ai-reference-models are comparing it to the libraries listed below
- A scalable inference server for models optimized with OpenVINO™☆719Updated this week
- Intel® Extension for TensorFlow*☆336Updated 2 weeks ago
- Inference Model Manager for Kubernetes☆46Updated 5 years ago
- Computation using data flow graphs for scalable machine learning☆67Updated this week
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆456Updated this week
- oneAPI Collective Communications Library (oneCCL)☆227Updated this week
- oneCCL Bindings for Pytorch*☆91Updated this week
- Triton Model Analyzer is a CLI tool to help with better understanding of the compute and memory requirements of the Triton Inference Serv…☆466Updated 3 weeks ago
- A multi-user, distributed computing environment for running DL model training experiments on Intel® Xeon® Scalable processor-based system…☆392Updated 10 months ago
- A Python package for extending the official PyTorch that can easily obtain performance on Intel platform☆1,812Updated this week
- Guide for building custom op for TensorFlow☆378Updated 2 years ago
- ☆57Updated 4 years ago
- OpenVINO™ integration with TensorFlow☆179Updated 9 months ago
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆991Updated this week
- Reference implementations of MLPerf™ inference benchmarks☆1,343Updated this week
- TensorFlow-nGraph bridge☆136Updated 4 years ago
- Issues related to MLPerf™ training policies, including rules and suggested changes☆94Updated 3 weeks ago
- A benchmark framework for Tensorflow☆1,150Updated last year
- To make it easy to benchmark AI accelerators☆183Updated 2 years ago
- TensorFlow/TensorRT integration☆739Updated last year
- SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX R…☆2,365Updated this week
- Reference models for Intel(R) Gaudi(R) AI Accelerator☆162Updated last month
- A performant and modular runtime for TensorFlow☆758Updated last month
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆249Updated this week
- High performance Cross-platform Inference-engine, you could run Anakin on x86-cpu,arm, nv-gpu, amd-gpu,bitmain and cambricon devices.☆533Updated 2 years ago
- A profiling and performance analysis tool for TensorFlow☆369Updated this week
- Representation and Reference Lowering of ONNX Models in MLIR Compiler Infrastructure☆834Updated last week
- A flexible and efficient deep neural network (DNN) compiler that generates high-performance executable from a DNN model description.☆980Updated 6 months ago
- Common utilities for ONNX converters☆261Updated 4 months ago
- Examples for using ONNX Runtime for model training.☆330Updated 5 months ago