HabanaAI / Setup_and_InstallLinks
Setup and Installation Instructions for Habana binaries, docker image creation
☆25Updated last month
Alternatives and similar repositories for Setup_and_Install
Users that are interested in Setup_and_Install are comparing it to the libraries listed below
Sorting:
- Large Language Model Text Generation Inference on Habana Gaudi☆34Updated 3 months ago
- Intel® Extension for DeepSpeed* is an extension to DeepSpeed that brings feature support with SYCL kernels on Intel GPU(XPU) device. Note…☆61Updated 2 weeks ago
- Tutorials for running models on First-gen Gaudi and Gaudi2 for Training and Inference. The source files for the tutorials on https://dev…☆62Updated 3 weeks ago
- Bandwidth test for ROCm☆60Updated this week
- Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)☆190Updated this week
- Development repository for the Triton language and compiler☆125Updated this week
- Reference models for Intel(R) Gaudi(R) AI Accelerator☆166Updated last week
- This repository contains Dockerfiles, scripts, yaml files, Helm charts, etc. used to scale out AI containers with versions of TensorFlow …☆48Updated last week
- ☆113Updated this week
- Machine Learning Agility (MLAgility) benchmark and benchmarking tools☆39Updated 2 months ago
- [DEPRECATED] Moved to ROCm/rocm-libraries repo☆110Updated this week
- ☆40Updated this week
- A high-throughput and memory-efficient inference and serving engine for LLMs☆77Updated this week
- AMD related optimizations for transformer models☆80Updated 3 weeks ago
- This repository hosts code that supports the testing infrastructure for the PyTorch organization. For example, this repo hosts the logic …☆96Updated this week
- AMD SMI☆78Updated this week
- DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.☆13Updated last month
- Accelerate your Gen AI with NVIDIA NIM and NVIDIA AI Workbench☆174Updated 2 months ago
- oneCCL Bindings for Pytorch*☆99Updated this week
- General policies for MLPerf™ including submission rules, coding standards, etc.☆29Updated this week
- ☆48Updated this week
- OpenAI Triton backend for Intel® GPUs☆193Updated this week
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆87Updated this week
- A collection of examples for the ROCm software stack☆228Updated this week
- Computation using data flow graphs for scalable machine learning☆68Updated this week
- GPU Stress Test is a tool to stress the compute engine of NVIDIA Tesla GPU’s by running a BLAS matrix multiply using different data types…☆103Updated last week
- ☆20Updated last week
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆303Updated this week
- AI Tensor Engine for ROCm☆232Updated this week
- A Fusion Code Generator for NVIDIA GPUs (commonly known as "nvFuser")☆343Updated this week