hidet-org / hidet
An open-source efficient deep learning framework/compiler, written in python.
☆668Updated this week
Alternatives and similar repositories for hidet:
Users that are interested in hidet are comparing it to the libraries listed below
- A Python-level JIT compiler designed to make unmodified PyTorch programs faster.☆1,016Updated 9 months ago
- A library to analyze PyTorch traces.☆323Updated last month
- Pipeline Parallelism for PyTorch☆736Updated 4 months ago
- Backward compatible ML compute opset inspired by HLO/MHLO☆428Updated this week
- Flash Attention in ~100 lines of CUDA (forward pass only)☆681Updated 2 weeks ago
- This repository contains the experimental PyTorch native float8 training UX☆219Updated 5 months ago
- A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.☆505Updated 2 months ago
- Mirage: Automatically Generating Fast GPU Kernels without Programming in Triton/CUDA☆714Updated this week
- Applied AI experiments and examples for PyTorch☆211Updated this week
- BitBLAS is a library to support mixed-precision matrix multiplications, especially for quantized LLM deployment.☆496Updated this week
- Fast low-bit matmul kernels in Triton☆187Updated last week
- A Fusion Code Generator for NVIDIA GPUs (commonly known as "nvFuser")☆291Updated this week
- Implementation of a Transformer, but completely in Triton☆251Updated 2 years ago
- The Torch-MLIR project aims to provide first class support from the PyTorch ecosystem to the MLIR ecosystem.☆1,401Updated this week
- ☆275Updated this week
- ☆170Updated this week
- Representation and Reference Lowering of ONNX Models in MLIR Compiler Infrastructure☆797Updated this week
- A library of GPU kernels for sparse matrix operations.☆251Updated 4 years ago
- A CPU+GPU Profiling library that provides access to timeline traces and hardware performance counters.☆755Updated last week
- Composable Kernel: Performance Portable Programming Model for Machine Learning Tensor Operators☆334Updated this week
- ☆154Updated 7 months ago
- FP16xINT4 LLM inference kernel that can achieve near-ideal ~4x speedups up to medium batchsizes of 16-32 tokens.☆680Updated 4 months ago
- The Tensor Algebra SuperOptimizer for Deep Learning☆696Updated last year
- ☆178Updated 6 months ago
- Fast CUDA matrix multiplication from scratch☆579Updated last year
- Shared Middle-Layer for Triton Compilation☆220Updated this week
- Repository for the QUIK project, enabling the use of 4bit kernels for generative inference - EMNLP 2024☆175Updated 9 months ago
- torch::deploy (multipy for non-torch uses) is a system that lets you get around the GIL problem by running multiple Python interpreters i…☆179Updated last month
- Collection of kernels written in Triton language☆90Updated 2 months ago
- Tile primitives for speedy kernels☆1,923Updated this week