intel / ai-reference-models
Intel® AI Reference Models: contains Intel optimizations for running deep learning workloads on Intel® Xeon® Scalable processors and Intel® Data Center GPUs
☆683Updated this week
Related projects ⓘ
Alternatives and complementary repositories for ai-reference-models
- A scalable inference server for models optimized with OpenVINO™☆675Updated this week
- TensorFlow/TensorRT integration☆736Updated 11 months ago
- Inference Model Manager for Kubernetes☆46Updated 5 years ago
- Intel® Extension for TensorFlow*☆318Updated last month
- Computation using data flow graphs for scalable machine learning☆67Updated this week
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆409Updated this week
- Reference implementations of MLPerf™ inference benchmarks☆1,238Updated this week
- A Python package for extending the official PyTorch that can easily obtain performance on Intel platform☆1,624Updated this week
- oneAPI Collective Communications Library (oneCCL)☆206Updated this week
- Actively maintained ONNX Optimizer☆647Updated 8 months ago
- SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX R…☆2,227Updated this week
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆943Updated this week
- Explainable AI Tooling (XAI). XAI is used to discover and explain a model's prediction in a way that is interpretable to the user. Releva…☆36Updated last month
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆152Updated this week
- To make it easy to benchmark AI accelerators☆179Updated last year
- A profiling and performance analysis tool for TensorFlow☆360Updated this week
- A multi-user, distributed computing environment for running DL model training experiments on Intel® Xeon® Scalable processor-based system…☆392Updated 6 months ago
- Triton Model Analyzer is a CLI tool to help with better understanding of the compute and memory requirements of the Triton Inference Serv…☆433Updated last week
- ROCm Communication Collectives Library (RCCL)☆268Updated this week
- Reference implementations of MLPerf™ training benchmarks☆1,617Updated last month
- ONNXMLTools enables conversion of models to ONNX☆1,024Updated 5 months ago
- Examples for using ONNX Runtime for model training.☆312Updated 3 weeks ago
- Common utilities for ONNX converters☆251Updated 5 months ago
- TorchBench is a collection of open source benchmarks used to evaluate PyTorch performance.☆875Updated this week
- Issues related to MLPerf™ training policies, including rules and suggested changes☆93Updated last month
- A performant and modular runtime for TensorFlow☆756Updated last month
- Convert TensorFlow, Keras, Tensorflow.js and Tflite models to ONNX☆2,327Updated 2 months ago
- ☆236Updated 3 years ago
- Explore the Capabilities of the TensorRT Platform☆260Updated 3 years ago
- ☆58Updated 4 years ago