intel / ai-reference-models
Intel® AI Reference Models: contains Intel optimizations for running deep learning workloads on Intel® Xeon® Scalable processors and Intel® Data Center GPUs
☆692Updated this week
Alternatives and similar repositories for ai-reference-models:
Users that are interested in ai-reference-models are comparing it to the libraries listed below
- A scalable inference server for models optimized with OpenVINO™☆701Updated this week
- Inference Model Manager for Kubernetes☆46Updated 5 years ago
- Intel® Extension for TensorFlow*☆329Updated 2 weeks ago
- Computation using data flow graphs for scalable machine learning☆67Updated this week
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆436Updated this week
- oneAPI Collective Communications Library (oneCCL)☆218Updated this week
- OpenVINO™ integration with TensorFlow☆179Updated 6 months ago
- A multi-user, distributed computing environment for running DL model training experiments on Intel® Xeon® Scalable processor-based system…☆392Updated 8 months ago
- SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX R…☆2,307Updated this week
- Reference implementations of MLPerf™ inference benchmarks☆1,289Updated this week
- Composable Kernel: Performance Portable Programming Model for Machine Learning Tensor Operators☆338Updated this week
- ONNX Script enables developers to naturally author ONNX functions and models using a subset of Python.☆312Updated this week
- A Python package for extending the official PyTorch that can easily obtain performance on Intel platform☆1,716Updated this week
- Issues related to MLPerf™ training policies, including rules and suggested changes☆94Updated 2 months ago
- A performant and modular runtime for TensorFlow☆759Updated this week
- Actively maintained ONNX Optimizer☆662Updated this week
- oneCCL Bindings for Pytorch*☆87Updated 3 weeks ago
- Examples for using ONNX Runtime for model training.☆324Updated 3 months ago
- Profiling Tools Interfaces for GPU (PTI for GPU) is a set of Getting Started Documentation and Tools Library to start performance analysi…☆215Updated 2 weeks ago
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆200Updated this week
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆968Updated this week
- A profiling and performance analysis tool for TensorFlow☆363Updated this week
- OpenAI Triton backend for Intel® GPUs☆156Updated this week
- Common utilities for ONNX converters☆257Updated last month
- ☆403Updated this week
- Intel® Extension for DeepSpeed* is an extension to DeepSpeed that brings feature support with SYCL kernels on Intel GPU(XPU) device. Note…☆59Updated last month
- To make it easy to benchmark AI accelerators☆183Updated 2 years ago
- ☆100Updated 2 months ago
- A Fusion Code Generator for NVIDIA GPUs (commonly known as "nvFuser")☆294Updated this week
- The Tensor Algebra SuperOptimizer for Deep Learning☆695Updated 2 years ago