sjtu-zhao-lab / SALOView external linksLinks
An efficient spatial accelerator enabling hybrid sparse attention mechanisms for long sequences
☆31Mar 7, 2024Updated last year
Alternatives and similar repositories for SALO
Users that are interested in SALO are comparing it to the libraries listed below
Sorting:
- ViTALiTy (HPCA'23) Code Repository☆23Mar 13, 2023Updated 2 years ago
- [HPCA'21] SpAtten: Efficient Sparse Attention Architecture with Cascade Token and Head Pruning☆122Aug 27, 2024Updated last year
- [TCAD'23] AccelTran: A Sparsity-Aware Accelerator for Transformers☆56Nov 22, 2023Updated 2 years ago
- [HPCA 2023] ViTCoD: Vision Transformer Acceleration via Dedicated Algorithm and Accelerator Co-Design☆127Jun 27, 2023Updated 2 years ago
- A co-design architecture on sparse attention☆55Aug 23, 2021Updated 4 years ago
- ☆48Aug 23, 2021Updated 4 years ago
- An analytical framework that models hardware dataflow of tensor applications on spatial architectures using the relation-centric notation…☆87Apr 28, 2024Updated last year
- A comprehensive content-addressable accelerator simulation framework.☆20Nov 15, 2024Updated last year
- An Optimizing Framework on MLIR for Efficient FPGA-based Accelerator Generation☆56Mar 22, 2024Updated last year
- ☆35Dec 22, 2025Updated last month
- This is my hobby project with System Verilog to accelerate LeViT Network which contain CNN and Attention layer.☆32Aug 13, 2024Updated last year
- You can run it on pynq z1. The repository contains the relevant Verilog code, Vivado configuration and C code for sdk testing. The size o…☆229Mar 24, 2024Updated last year
- ☆10Mar 8, 2025Updated 11 months ago
- ☆12Aug 18, 2023Updated 2 years ago
- NeuPIMs: NPU-PIM Heterogeneous Acceleration for Batched LLM Inferencing☆108Jun 19, 2024Updated last year
- HW/SW co-design of sentence-level energy optimizations for latency-aware multi-task NLP inference☆54Mar 24, 2024Updated last year
- ☆49Apr 22, 2021Updated 4 years ago
- bitfusion verilog implementation☆12Feb 21, 2022Updated 3 years ago
- [HPCA 2026 Best Paper Candidate] Official implementation of "Focus: A Streaming Concentration Architecture for Efficient Vision-Language …☆30Feb 8, 2026Updated last week
- RISC-V-based many-core neuromorphic architecture☆14Aug 3, 2025Updated 6 months ago
- ☆46Apr 8, 2023Updated 2 years ago
- Edge-MoE: Memory-Efficient Multi-Task Vision Transformer Architecture with Task-level Sparsity via Mixture-of-Experts☆132May 10, 2024Updated last year
- FPGA-based hardware accelerator for Vision Transformer (ViT), with Hybrid-Grained Pipeline.☆125Jan 20, 2025Updated last year
- Lab code for three-day lecture, "Designing CNN Accelerators using Bluespec System Verilog", given at SNU in December 2017☆32Sep 22, 2018Updated 7 years ago
- An FPGA Accelerator for Transformer Inference☆93Apr 29, 2022Updated 3 years ago
- Implementation of paper "GraphACT: Accelerating GCN Training on CPU-FPGA Heterogeneous Platform".☆12Jun 25, 2020Updated 5 years ago
- DATE'24 paper: "Hierarchical Source-to-Post-Route QoR Prediction in High-Level Synthesis with GNNs"☆19Dec 10, 2024Updated last year
- SSR: Spatial Sequential Hybrid Architecture for Latency Throughput Tradeoff in Transformer Acceleration (Full Paper Accepted in FPGA'24)☆35Updated this week
- This repository presents the source code for the paper "MILLION: Mastering Long-Context LLM Inference Via Outlier-Immunized KV Product Qu…☆23Apr 2, 2025Updated 10 months ago
- NATSA is the first near-data-processing accelerator for time series analysis based on the Matrix Profile (SCRIMP) algorithm. NATSA exploi…☆16Jun 14, 2023Updated 2 years ago
- TinyVers Heterogeneous SoC consists of a reconfigurable FlexML accelerator, a RISC-V processor, an eMRAM and a power management system.☆23Jul 12, 2023Updated 2 years ago
- LoAS: Fully Temporal-Parallel Dataflow for Dual-Sparse Spiking Neural Networks, MICRO 2024.☆17Mar 19, 2025Updated 10 months ago
- H2-LLM: Hardware-Dataflow Co-Exploration for Heterogeneous Hybrid-Bonding-based Low-Batch LLM Inference☆87Apr 26, 2025Updated 9 months ago
- ☆37Jan 20, 2022Updated 4 years ago
- ☆113Nov 17, 2023Updated 2 years ago
- The official implementation of the DAC 2024 paper GQA-LUT☆20Dec 20, 2024Updated last year
- An end-to-end GCN inference accelerator written in HLS☆18Apr 5, 2022Updated 3 years ago
- The framework for the paper "Inter-layer Scheduling Space Definition and Exploration for Tiled Accelerators" in ISCA 2023.☆82Mar 12, 2025Updated 11 months ago
- SDA: Low-Bit Stable Diffusion Acceleration on Edge FPGAs☆18May 23, 2024Updated last year