tenstorrent / tt-inference-serverLinks
☆20Updated this week
Alternatives and similar repositories for tt-inference-server
Users that are interested in tt-inference-server are comparing it to the libraries listed below
Sorting:
- The TT-Forge FE is a graph compiler designed to optimize and transform computational graphs for deep learning models, enhancing their per…☆46Updated this week
- Tenstorrent MLIR compiler☆151Updated this week
- Tenstorrent TT-BUDA Repository☆314Updated 3 months ago
- Buda Compiler Backend for Tenstorrent devices☆29Updated 3 months ago
- TVM for Tenstorrent ASICs☆23Updated last week
- Tenstorrent Kernel Module☆46Updated this week
- Tenstorrent console based hardware information program☆47Updated last week
- ☆29Updated 4 months ago
- Tenstorrent Firmware repository☆15Updated 2 weeks ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆18Updated this week
- Tenstorrent's MLIR Based Compiler. We aim to enable developers to run AI on all configurations of Tenstorrent hardware, through an open-s…☆80Updated this week
- QuickReduce is a performant all-reduce library designed for AMD ROCm that supports inline compression.☆30Updated 4 months ago
- IREE plugin repository for the AMD AIE accelerator☆98Updated this week
- ☆102Updated this week
- Tenstorrent Topology (TT-Topology) is a command line utility used to flash multiple NB cards on a system to use specific eth routing conf…☆11Updated this week
- Machine-Learning Accelerator System Exploration Tools☆172Updated last month
- IREE's PyTorch Frontend, based on Torch Dynamo.☆93Updated this week
- Attention in SRAM on Tenstorrent Grayskull☆36Updated last year
- Repo for AI Compiler team. The intended purpose of this repo is for implementation of a PJRT device.☆18Updated this week
- Development repository for the Triton language and compiler☆125Updated this week
- Fork of LLVM to support AMD AIEngine processors☆152Updated this week
- PyTorch emulation library for Microscaling (MX)-compatible data formats☆257Updated last month
- OpenAI Triton backend for Intel® GPUs☆193Updated this week
- Unofficial description of the CUDA assembly (SASS) instruction sets.☆107Updated this week
- ☆28Updated 2 years ago
- ⭐️ TTNN Compiler for PyTorch 2 ⭐️ Enables running PyTorch models on Tenstorrent hardware using eager or compile path☆52Updated this week
- An experimental CPU backend for Triton☆135Updated last month
- ASPLOS'24: Optimal Kernel Orchestration for Tensor Programs with Korch☆37Updated 3 months ago
- ☆143Updated last year
- ☆48Updated last week