intel / ai-reference-models
Intel® AI Reference Models: contains Intel optimizations for running deep learning workloads on Intel® Xeon® Scalable processors and Intel® Data Center GPUs
☆674Updated this week
Related projects: ⓘ
- A scalable inference server for models optimized with OpenVINO™☆657Updated this week
- TensorFlow/TensorRT integration☆737Updated 9 months ago
- Reference implementations of MLPerf™ inference benchmarks☆1,188Updated 2 weeks ago
- Inference Model Manager for Kubernetes☆46Updated 5 years ago
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆380Updated this week
- Intel® Extension for TensorFlow*☆314Updated 2 weeks ago
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆906Updated this week
- Examples for using ONNX Runtime for model training.☆301Updated last month
- A profiling and performance analysis tool for TensorFlow☆355Updated this week
- Representation and Reference Lowering of ONNX Models in MLIR Compiler Infrastructure☆743Updated this week
- Actively maintained ONNX Optimizer☆634Updated 6 months ago
- A performant and modular runtime for TensorFlow☆754Updated last month
- Computation using data flow graphs for scalable machine learning☆65Updated this week
- A Python package for extending the official PyTorch that can easily obtain performance on Intel platform☆1,554Updated this week
- OpenVINO™ integration with TensorFlow☆178Updated 2 months ago
- Reference implementations of MLPerf™ training benchmarks☆1,597Updated last month
- A multi-user, distributed computing environment for running DL model training experiments on Intel® Xeon® Scalable processor-based system…☆393Updated 4 months ago
- ☆392Updated last week
- Issues related to MLPerf™ training policies, including rules and suggested changes☆92Updated last week
- FB (Facebook) + GEMM (General Matrix-Matrix Multiplication) - https://code.fb.com/ml-applications/fbgemm/☆1,169Updated this week
- SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX R…☆2,152Updated this week
- ONNXMLTools enables conversion of models to ONNX☆992Updated 3 months ago
- oneAPI Collective Communications Library (oneCCL)☆189Updated 3 weeks ago
- Convert tf.keras/Keras models to ONNX☆379Updated 3 years ago
- nGraph has moved to OpenVINO☆1,355Updated 3 years ago
- Triton Model Analyzer is a CLI tool to help with better understanding of the compute and memory requirements of the Triton Inference Serv…☆419Updated 2 weeks ago
- Run Generative AI models using native OpenVINO C++ API☆107Updated this week
- A benchmark framework for Tensorflow☆1,144Updated 11 months ago
- Tensorflow Backend for ONNX☆1,272Updated 5 months ago
- The Torch-MLIR project aims to provide first class support from the PyTorch ecosystem to the MLIR ecosystem.☆1,301Updated this week