A Python package for extending the official PyTorch that can easily obtain performance on Intel platform
☆2,010Mar 30, 2026Updated 2 weeks ago
Alternatives and similar repositories for intel-extension-for-pytorch
Users that are interested in intel-extension-for-pytorch are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- Intel® Extension for TensorFlow*☆352Oct 29, 2025Updated 5 months ago
- ⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Pl…☆2,178Oct 8, 2024Updated last year
- SOTA low-bit LLM quantization (INT8/FP8/MXFP8/INT4/MXFP4/NVFP4) & sparsity; leading model compression techniques on PyTorch, TensorFlow, …☆2,612Updated this week
- oneCCL Bindings for Pytorch* (deprecated)☆104Dec 31, 2025Updated 3 months ago
- OpenAI Triton backend for Intel® GPUs☆241Updated this week
- GPU virtual machines on DigitalOcean Gradient AI • AdGet to production fast with high-performance AMD and NVIDIA GPUs you can spin up in seconds. The definition of operational simplicity.
- oneAPI Deep Neural Network Library (oneDNN)☆3,974Updated this week
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆561Apr 2, 2026Updated last week
- Intel® Extension for DeepSpeed* is an extension to DeepSpeed that brings feature support with SYCL kernels on Intel GPU(XPU) device. Note…☆65Jun 30, 2025Updated 9 months ago
- ☆83Updated this week
- Intel® AI Reference Models: contains Intel optimizations for running deep learning workloads on Intel® Xeon® Scalable processors and Inte…☆731Feb 11, 2026Updated 2 months ago
- Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, DeepSeek, Mixtral, Gemma, Phi, MiniCPM, Qwen-VL, MiniCPM-V,…☆8,752Jan 28, 2026Updated 2 months ago
- oneAPI Collective Communications Library (oneCCL)☆259Updated this week
- OpenVINO™ is an open source toolkit for optimizing and deploying AI inference☆10,054Updated this week
- A Python-level JIT compiler designed to make unmodified PyTorch programs faster.☆1,078Apr 17, 2024Updated last year
- Managed hosting for WordPress and PHP on Cloudways • AdManaged hosting for WordPress, Magento, Laravel, or PHP apps, on multiple cloud providers. Deploy in minutes on Cloudways by DigitalOcean.
- Intel® NPU Acceleration Library☆709Apr 24, 2025Updated 11 months ago
- Development repository for the Triton language and compiler☆18,902Updated this week
- PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT☆2,963Updated this week
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆486Updated this week
- SYCL* Templates for Linear Algebra (SYCL*TLA) - SYCL based CUTLASS implementation for Intel GPUs☆73Updated this week
- ☆160Mar 12, 2026Updated last month
- Transformer related optimization, including BERT, GPT☆6,412Mar 27, 2024Updated 2 years ago
- Accessible large language models via k-bit quantization for PyTorch.☆8,107Updated this week
- Fast and memory-efficient exact attention☆23,185Apr 6, 2026Updated last week
- Deploy open-source AI quickly and easily - Bonus Offer • AdRunpod Hub is built for open source. One-click deployment and autoscaling endpoints without provisioning your own infrastructure.
- AI PC starter app for doing AI image creation, image stylizing, and chatbot on a PC powered by an Intel® Arc™ GPU.☆823Updated this week
- 🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization…☆3,354Apr 2, 2026Updated last week
- PyTorch extensions for high performance and large scale training.☆3,405Apr 26, 2025Updated 11 months ago
- ☆436Sep 18, 2025Updated 6 months ago
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆1,147Updated this week
- Intel® Graphics Compute Runtime for oneAPI Level Zero and OpenCL™ Driver☆1,364Updated this week
- Profiling Tools Interfaces for GPU (PTI for GPU) is a set of Getting Started Documentation and Tools Library to start performance analysi…☆264Updated this week
- SOTA rounding-based quantization for high-accuracy low-bit LLM inference, seamlessly optimized for CPU/XPU/CUDA, with multi-datatype supp…☆957Updated this week
- FB (Facebook) + GEMM (General Matrix-Matrix Multiplication) - https://code.fb.com/ml-applications/fbgemm/☆1,552Updated this week
- Virtual machines for every use case on DigitalOcean • AdGet dependable uptime with 99.99% SLA, simple security tools, and predictable monthly pricing with DigitalOcean's virtual machines, called Droplets.
- Olive: Simplify ML Model Finetuning, Conversion, Quantization, and Optimization for CPUs, GPUs and NPUs.☆2,289Updated this week
- ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator☆19,779Apr 7, 2026Updated last week
- CUDA Templates and Python DSLs for High-Performance Linear Algebra☆9,564Updated this week
- Intel staging area for llvm.org contribution. Home for Intel LLVM-based projects.☆1,451Updated this week
- The Torch-MLIR project aims to provide first class support from the PyTorch ecosystem to the MLIR ecosystem.☆1,781Apr 7, 2026Updated last week
- PyTorch native quantization and sparsity for training and inference☆2,769Updated this week
- A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit and 4-bit floating point (FP8 and FP4) precision on H…☆3,269Updated this week