FB (Facebook) + GEMM (General Matrix-Matrix Multiplication) - https://code.fb.com/ml-applications/fbgemm/
☆1,534Updated this week
Alternatives and similar repositories for FBGEMM
Users that are interested in FBGEMM are comparing it to the libraries listed below
Sorting:
- CUDA Templates and Python DSLs for High-Performance Linear Algebra☆9,315Updated this week
- Low-precision matrix multiplication☆1,831Jan 29, 2024Updated 2 years ago
- Compiler for Neural Network hardware accelerators☆3,326May 11, 2024Updated last year
- Quantized Neural Network PACKage - mobile-optimized implementation of quantized neural network operators☆1,546Aug 28, 2019Updated 6 years ago
- A flexible and efficient deep neural network (DNN) compiler that generates high-performance executable from a DNN model description.☆1,006Sep 19, 2024Updated last year
- A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit and 4-bit floating point (FP8 and FP4) precision on H…☆3,170Feb 21, 2026Updated last week
- oneAPI Deep Neural Network Library (oneDNN)☆3,956Updated this week
- Development repository for the Triton language and compiler☆18,460Feb 22, 2026Updated last week
- ☆1,992Jul 29, 2023Updated 2 years ago
- Transformer related optimization, including BERT, GPT☆6,394Mar 27, 2024Updated last year
- Open Machine Learning Compiler Framework☆13,142Updated this week
- A CPU+GPU Profiling library that provides access to timeline traces and hardware performance counters.☆922Updated this week
- PyTorch extensions for high performance and large scale training.☆3,400Apr 26, 2025Updated 10 months ago
- The Torch-MLIR project aims to provide first class support from the PyTorch ecosystem to the MLIR ecosystem.☆1,754Updated this week
- HugeCTR is a high efficiency GPU framework designed for Click-Through-Rate (CTR) estimating training☆1,045Sep 15, 2025Updated 5 months ago
- AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (N…☆4,706Jan 12, 2026Updated last month
- Pytorch domain library for recommendation systems☆2,471Updated this week
- A PyTorch Extension: Tools for easy mixed precision and distributed training in Pytorch☆8,926Updated this week
- Collective communications library with various primitives for multi-machine training.☆1,399Feb 12, 2026Updated 2 weeks ago
- The Tensor Algebra SuperOptimizer for Deep Learning☆739Jan 26, 2023Updated 3 years ago
- Optimized primitives for collective multi-GPU communication☆4,474Updated this week
- A Python-level JIT compiler designed to make unmodified PyTorch programs faster.☆1,075Apr 17, 2024Updated last year
- High-efficiency floating-point neural network inference operators for mobile, server, and Web☆2,263Updated this week
- FlashInfer: Kernel Library for LLM Serving☆5,009Updated this week
- common in-memory tensor structure☆1,169Jan 26, 2026Updated last month
- Enabling PyTorch on XLA Devices (e.g. Google TPU)☆2,751Dec 18, 2025Updated 2 months ago
- A tensor-aware point-to-point communication primitive for machine learning☆283Dec 17, 2025Updated 2 months ago
- A high performance and generic framework for distributed DNN training☆3,716Oct 3, 2023Updated 2 years ago
- Library for specialized dense and sparse matrix operations, and deep learning primitives.☆938Feb 14, 2026Updated 2 weeks ago
- Automatically Discovering Fast Parallelization Strategies for Distributed Deep Neural Network Training☆1,861Feb 20, 2026Updated last week
- ☆321Feb 17, 2026Updated last week
- Distributed Compiler based on Triton for Parallel Systems☆1,361Feb 13, 2026Updated 2 weeks ago
- PyTorch native quantization and sparsity for training and inference☆2,696Feb 22, 2026Updated last week
- [ARCHIVED] Cooperative primitives for CUDA C++. See https://github.com/NVIDIA/cccl☆1,820Oct 9, 2023Updated 2 years ago
- Tile primitives for speedy kernels☆3,183Updated this week
- BladeDISC is an end-to-end DynamIc Shape Compiler project for machine learning workloads.☆918Dec 30, 2024Updated last year
- A fast communication-overlapping library for tensor/expert parallelism on GPUs.☆1,261Aug 28, 2025Updated 6 months ago
- Ongoing research training transformer models at scale☆15,242Feb 21, 2026Updated last week
- A GPU-accelerated library containing highly optimized building blocks and an execution engine for data processing to accelerate deep lear…☆5,634Feb 19, 2026Updated last week