Richielee630 / TMMALinks
TMMA: A Tiled Matrix Multiplication Accelerator for Self-Attention Projections in Transformer Models, optimized for edge deployment on Xilinx KV260.
☆23Updated 7 months ago
Alternatives and similar repositories for TMMA
Users that are interested in TMMA are comparing it to the libraries listed below
Sorting:
- [HPCA 2023] ViTCoD: Vision Transformer Acceleration via Dedicated Algorithm and Accelerator Co-Design☆122Updated 2 years ago
- [HPCA'21] SpAtten: Efficient Sparse Attention Architecture with Cascade Token and Head Pruning☆111Updated last year
- SSR: Spatial Sequential Hybrid Architecture for Latency Throughput Tradeoff in Transformer Acceleration (Full Paper Accepted in FPGA'24)☆33Updated this week
- FPGA based Vision Transformer accelerator (Harvard CS205)☆137Updated 8 months ago
- ☆10Updated last year
- ViTALiTy (HPCA'23) Code Repository☆23Updated 2 years ago
- An FPGA Accelerator for Transformer Inference☆91Updated 3 years ago
- [TCAD'23] AccelTran: A Sparsity-Aware Accelerator for Transformers☆54Updated last year
- ☆12Updated 2 years ago
- Includes the SVD-based approximation algorithms for compressing deep learning models and the FPGA accelerators exploiting such approximat…☆16Updated 2 years ago
- FPGA-based hardware accelerator for Vision Transformer (ViT), with Hybrid-Grained Pipeline.☆98Updated 9 months ago
- Edge-MoE: Memory-Efficient Multi-Task Vision Transformer Architecture with Task-level Sparsity via Mixture-of-Experts☆127Updated last year
- Simulator for LLM inference on an abstract 3D AIMC-based accelerator☆24Updated last month
- An open source Verilog Based LeNet-1 Parallel CNNs Accelerator for FPGAs in Vivado 2017☆19Updated 6 years ago
- H2-LLM: Hardware-Dataflow Co-Exploration for Heterogeneous Hybrid-Bonding-based Low-Batch LLM Inference☆75Updated 6 months ago
- The official implementation of the DAC 2024 paper GQA-LUT☆20Updated 10 months ago
- Serpens is an HBM FPGA accelerator for SpMV☆22Updated last year
- [ASP-DAC 2025] "NeuronQuant: Accurate and Efficient Post-Training Quantization for Spiking Neural Networks" Official Implementation☆13Updated 8 months ago
- An efficient spatial accelerator enabling hybrid sparse attention mechanisms for long sequences☆30Updated last year
- ☆47Updated 4 years ago
- This is my hobby project with System Verilog to accelerate LeViT Network which contain CNN and Attention layer.☆23Updated last year
- (Verilog) A simple convolution layer implementation with systolic array structure☆13Updated 3 years ago
- A Reconfigurable Accelerator with Data Reordering Support for Low-Cost On-Chip Dataflow Switching☆68Updated last week
- Open-source of MSD framework☆16Updated 2 years ago
- ☆51Updated 3 months ago
- ☆70Updated last month
- ☆30Updated 7 months ago
- [DATE 2025] Official implementation and dataset of AIrchitect v2: Learning the Hardware Accelerator Design Space through Unified Represen…☆17Updated 9 months ago
- PALM: A Efficient Performance Simulator for Tiled Accelerators with Large-scale Model Training☆19Updated last year
- tpu-systolic-array-weight-stationary☆24Updated 4 years ago