pnnl / TCBNNLinks
☆34Updated 3 years ago
Alternatives and similar repositories for TCBNN
Users that are interested in TCBNN are comparing it to the libraries listed below
Sorting:
- Singular Binarized Neural Network based on GPU Bit Operations (see our SC-19 paper)☆15Updated 4 years ago
- ☆19Updated 4 years ago
- ☆39Updated 5 years ago
- Official implementation of "Searching for Winograd-aware Quantized Networks" (MLSys'20)☆27Updated last year
- Benchmark for matrix multiplications between dense and block sparse (BSR) matrix in TVM, blocksparse (Gray et al.) and cuSparse.☆23Updated 5 years ago
- Multi-target compiler for Sum-Product Networks, based on MLIR and LLVM.☆24Updated 9 months ago
- GoldenEye is a functional simulator with fault injection capabilities for common and emerging numerical formats, implemented for the PyTo…☆26Updated 11 months ago
- Automatic Mapping Generation, Verification, and Exploration for ISA-based Spatial Accelerators☆115Updated 2 years ago
- SparseTIR: Sparse Tensor Compiler for Deep Learning☆138Updated 2 years ago
- ☆27Updated 5 years ago
- PyTorch extension for emulating FP8 data formats on standard FP32 Xeon/GPU hardware.☆111Updated 9 months ago
- Automatic Schedule Exploration and Optimization Framework for Tensor Computations☆180Updated 3 years ago
- Artifacts of EVT ASPLOS'24☆26Updated last year
- ☆14Updated 4 years ago
- ☆32Updated 3 years ago
- Magicube is a high-performance library for quantized sparse matrix operations (SpMM and SDDMM) of deep learning on Tensor Cores.☆89Updated 2 years ago
- ☆19Updated 3 years ago
- ☆40Updated 3 years ago
- ☆108Updated last year
- Benchmark PyTorch Custom Operators☆14Updated 2 years ago
- Simulator for BitFusion☆101Updated 5 years ago
- ☆45Updated 4 years ago
- Chameleon: Adaptive Code Optimization for Expedited Deep Neural Network Compilation☆27Updated 5 years ago
- Official implementation of EMNLP'23 paper "Revisiting Block-based Quantisation: What is Important for Sub-8-bit LLM Inference?"☆23Updated last year
- BSQ: Exploring Bit-Level Sparsity for Mixed-Precision Neural Network Quantization (ICLR 2021)☆42Updated 4 years ago
- GEMM and Winograd based convolutions using CUTLASS☆27Updated 5 years ago
- CUDA templates for tile-sparse matrix multiplication based on CUTLASS.☆50Updated 7 years ago
- ☆23Updated 2 years ago
- ☆31Updated 3 years ago
- HW/SW co-design of sentence-level energy optimizations for latency-aware multi-task NLP inference☆52Updated last year