HabanaAI / DeepSpeedLinks
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
☆13Updated last month
Alternatives and similar repositories for DeepSpeed
Users that are interested in DeepSpeed are comparing it to the libraries listed below
Sorting:
- RCCL Performance Benchmark Tests☆70Updated this week
- Development repository for the Triton language and compiler☆125Updated this week
- oneCCL Bindings for Pytorch*☆99Updated this week
- A high-throughput and memory-efficient inference and serving engine for LLMs☆77Updated this week
- A CUTLASS implementation using SYCL☆30Updated last week
- oneAPI Collective Communications Library (oneCCL)☆238Updated last week
- ☆40Updated this week
- Intel® Extension for DeepSpeed* is an extension to DeepSpeed that brings feature support with SYCL kernels on Intel GPU(XPU) device. Note…☆61Updated 2 weeks ago
- ☆48Updated this week
- A Fusion Code Generator for NVIDIA GPUs (commonly known as "nvFuser")☆343Updated this week
- OpenAI Triton backend for Intel® GPUs☆193Updated this week
- An extension library of WMMA API (Tensor Core API)☆99Updated last year
- ☆62Updated 7 months ago
- ROCm Communication Collectives Library (RCCL)☆349Updated this week
- Composable Kernel: Performance Portable Programming Model for Machine Learning Tensor Operators☆437Updated this week
- Reference models for Intel(R) Gaudi(R) AI Accelerator☆166Updated last week
- Provides the examples to write and build Habana custom kernels using the HabanaTools☆22Updated 3 months ago
- Intel® Tensor Processing Primitives extension for Pytorch*☆17Updated this week
- PArametrized Recommendation and Ai Model benchmark is a repository for development of numerous uBenchmarks as well as end to end nets for…☆147Updated 2 weeks ago
- Training material for Nsight developer tools☆161Updated 11 months ago
- rocSHMEM intra-kernel networking runtime for AMD dGPUs on the ROCm platform.☆91Updated this week
- Experimental projects related to TensorRT☆107Updated this week
- ☆148Updated this week
- ☆216Updated last year
- QuickReduce is a performant all-reduce library designed for AMD ROCm that supports inline compression.☆30Updated 4 months ago
- AI Accelerator Benchmark focuses on evaluating AI Accelerators from a practical production perspective, including the ease of use and ver…☆251Updated 2 weeks ago
- PArallelLOOPgEneratoR: Threaded Loops Code Generation Infrastructure targeting Tensor Contraction Applications such as GEMMs, Convolution…☆19Updated last month
- ☆25Updated 3 weeks ago
- Matrix Multiply-Accumulate with CUDA and WMMA( Tensor Core)☆138Updated 4 years ago
- An efficient GPU support for LLM inference with x-bit quantization (e.g. FP6,FP5).☆255Updated 8 months ago