tenstorrent / tt-inference-serverLinks
☆32Updated this week
Alternatives and similar repositories for tt-inference-server
Users that are interested in tt-inference-server are comparing it to the libraries listed below
Sorting:
- The TT-Forge FE is a graph compiler designed to optimize and transform computational graphs for deep learning models, enhancing their per…☆51Updated this week
- Tenstorrent's MLIR Based Compiler. We aim to enable developers to run AI on all configurations of Tenstorrent hardware, through an open-s…☆135Updated this week
- TT-Studio : An all-in-one platform to deploy and manage AI models optimized for Tenstorrent hardware with dedicated front-end demo applic…☆39Updated last week
- QuickReduce is a performant all-reduce library designed for AMD ROCm that supports inline compression.☆35Updated 2 months ago
- Tenstorrent TT-BUDA Repository☆315Updated 7 months ago
- Attention in SRAM on Tenstorrent Grayskull☆38Updated last year
- Tenstorrent MLIR compiler☆206Updated this week
- Tenstorrent console based hardware information program☆54Updated this week
- Tenstorrent Kernel Module☆55Updated this week
- AI Tensor Engine for ROCm☆296Updated this week
- ☆28Updated 7 months ago
- ⭐️ TTNN Compiler for PyTorch 2 ⭐️ Enables running PyTorch models on Tenstorrent hardware using eager or compile path☆60Updated this week
- Repo for AI Compiler team. The intended purpose of this repo is for implementation of a PJRT device.☆40Updated this week
- Evaluating Large Language Models for CUDA Code Generation ComputeEval is a framework designed to generate and evaluate CUDA code from Lar…☆70Updated last month
- TT-NN operator library, and TT-Metalium low level kernel programming model.☆1,250Updated this week
- RCCL Performance Benchmark Tests☆78Updated last week
- TVM for Tenstorrent ASICs☆27Updated last month
- OpenAI Triton backend for Intel® GPUs☆219Updated this week
- An experimental CPU backend for Triton☆155Updated last week
- GPUOcelot: A dynamic compilation framework for PTX☆211Updated 8 months ago
- Buda Compiler Backend for Tenstorrent devices☆30Updated 7 months ago
- Efficient implementation of DeepSeek Ops (Blockwise FP8 GEMM, MoE, and MLA) for AMD Instinct MI300X☆71Updated this week
- ☆76Updated this week
- SHARK Inference Modeling and Serving☆53Updated this week
- ☆72Updated 8 months ago
- Custom PTX Instruction Benchmark☆131Updated 8 months ago
- KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA (+ more DSLs)☆642Updated this week
- IREE's PyTorch Frontend, based on Torch Dynamo.☆99Updated this week
- A high-throughput and memory-efficient inference and serving engine for LLMs☆24Updated this week
- Tilus is a tile-level kernel programming language with explicit control over shared memory and registers.☆394Updated this week