google-research-datasets / tpu_graphs
☆124Updated 7 months ago
Alternatives and similar repositories for tpu_graphs:
Users that are interested in tpu_graphs are comparing it to the libraries listed below
- ☆77Updated last year
- Cataloging released Triton kernels.☆168Updated last month
- Collection of kernels written in Triton language☆105Updated this week
- Flash-LLM: Enabling Cost-Effective and Highly-Efficient Large Generative Model Inference with Unstructured Sparsity☆199Updated last year
- Hydragen: High-Throughput LLM Inference with Shared Prefixes☆34Updated 9 months ago
- Small scale distributed training of sequential deep learning models, built on Numpy and MPI.☆118Updated last year
- Official repository of Sparse ISO-FLOP Transformations for Maximizing Training Efficiency☆25Updated 6 months ago
- ☆100Updated 5 months ago
- WholeGraph - large scale Graph Neural Networks☆101Updated 2 months ago
- Fast and memory-efficient exact attention☆58Updated this week
- ☆246Updated 6 months ago
- Experiments on Multi-Head Latent Attention☆67Updated 6 months ago
- PyTorch-Direct code on top of PyTorch-1.8.0nightly (e152ca5) for Large Graph Convolutional Network Training with GPU-Oriented Data Commun…☆45Updated last year
- Fast Matrix Multiplications for Lookup Table-Quantized LLMs☆229Updated this week
- extensible collectives library in triton☆83Updated 4 months ago
- Memory Optimizations for Deep Learning (ICML 2023)☆62Updated 11 months ago
- ☆23Updated last year
- Fast low-bit matmul kernels in Triton☆236Updated this week
- NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day☆255Updated last year
- Fast Hadamard transform in CUDA, with a PyTorch interface☆143Updated 8 months ago
- Cold Compress is a hackable, lightweight, and open-source toolkit for creating and benchmarking cache compression methods built on top of…☆117Updated 6 months ago
- [ICLR'25] Fast Inference of MoE Models with CPU-GPU Orchestration☆193Updated 3 months ago
- Explorations into some recent techniques surrounding speculative decoding☆240Updated last month
- The simplest implementation of recent Sparse Attention patterns for efficient LLM inference.☆57Updated 3 weeks ago
- ☆24Updated last year
- A Python library transfers PyTorch tensors between CPU and NVMe☆104Updated 2 months ago
- KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems☆183Updated this week
- Boosting 4-bit inference kernels with 2:4 Sparsity☆64Updated 5 months ago
- A performant, memory-efficient checkpointing library for PyTorch applications, designed with large, complex distributed workloads in mind…☆154Updated 2 months ago
- [NeurIPS'23] Speculative Decoding with Big Little Decoder☆88Updated last year