☆35Dec 22, 2025Updated 2 months ago
Alternatives and similar repositories for quantized-training
Users that are interested in quantized-training are comparing it to the libraries listed below
Sorting:
- An efficient spatial accelerator enabling hybrid sparse attention mechanisms for long sequences☆31Mar 7, 2024Updated last year
- [HPCA 2023] ViTCoD: Vision Transformer Acceleration via Dedicated Algorithm and Accelerator Co-Design☆128Jun 27, 2023Updated 2 years ago
- Tender: Accelerating Large Language Models via Tensor Decompostion and Runtime Requantization (ISCA'24)☆31Jul 4, 2024Updated last year
- ☆143Jul 19, 2025Updated 7 months ago
- ViTALiTy (HPCA'23) Code Repository☆23Mar 13, 2023Updated 2 years ago
- Implementation of Microscaling data formats in SystemVerilog.☆29Jul 6, 2025Updated 8 months ago
- ☆11Aug 2, 2024Updated last year
- First Latency-Aware Competitive LLM Agent Benchmark☆26Jun 3, 2025Updated 9 months ago
- Here are some implementations of basic hardware units in RTL language (verilog for now), which can be used for area/power evaluation and …☆14Aug 25, 2023Updated 2 years ago
- LLM Inference with Microscaling Format☆34Nov 12, 2024Updated last year
- Official implementation for "Pruning Large Language Models with Semi-Structural Adaptive Sparse Training" (AAAI 2025)☆18Jul 1, 2025Updated 8 months ago
- HALO: Hadamard-Assisted Low-Precision Optimization and Training method for finetuning LLMs. 🚀 The official implementation of https://arx…☆29Feb 17, 2025Updated last year
- Torch2Chip (MLSys, 2024)☆55Apr 2, 2025Updated 11 months ago
- Low Precision Arithmetic Simulation in PyTorch - extension for posit and beyond☆16Dec 9, 2025Updated 2 months ago
- softfloat and softposit in Python☆15Aug 2, 2019Updated 6 years ago
- [TCAD'23] AccelTran: A Sparsity-Aware Accelerator for Transformers☆58Nov 22, 2023Updated 2 years ago
- [ICML 2024] When Linear Attention Meets Autoregressive Decoding: Towards More Effective and Efficient Linearized Large Language Models☆35Jun 12, 2024Updated last year
- ☆48Aug 23, 2021Updated 4 years ago
- PyTorchSim is a Comprehensive, Fast, and Accurate NPU Simulation Framework☆93Updated this week
- ☆32Oct 21, 2025Updated 4 months ago
- ☆20Dec 5, 2024Updated last year
- A Reconfigurable Accelerator with Data Reordering Support for Low-Cost On-Chip Dataflow Switching☆77Feb 26, 2026Updated last week
- DOSA: Differentiable Model-Based One-Loop Search for DNN Accelerators☆19Oct 10, 2024Updated last year
- ☆20Feb 12, 2025Updated last year
- ☆116Nov 17, 2023Updated 2 years ago
- ☆16Nov 14, 2022Updated 3 years ago
- ☆19Jul 30, 2024Updated last year
- Training with Block Minifloat number representation☆18May 2, 2021Updated 4 years ago
- A systolic array simulator for multi-cycle MACs and varying-byte words, with the paper accepted to HPCA 2022.☆85Nov 7, 2021Updated 4 years ago
- [NeurIPS 2023] Token-Scaled Logit Distillation for Ternary Weight Generative Language Models☆18Dec 6, 2023Updated 2 years ago
- ☆34Aug 27, 2025Updated 6 months ago
- Official Pytorch Implementation of "Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity"☆81Jul 7, 2025Updated 8 months ago
- PositNN - Framework for training and inference with neural nets usings posits☆20Jan 22, 2022Updated 4 years ago
- An analytical cost model evaluating DNN mappings (dataflows and tiling).☆247Apr 15, 2024Updated last year
- Serpens is an HBM FPGA accelerator for SpMV☆22Jul 26, 2024Updated last year
- FBI-LLM: Scaling Up Fully Binarized LLMs from Scratch via Autoregressive Distillation☆51Aug 24, 2025Updated 6 months ago
- FireQ: Fast INT4-FP8 Kernel and RoPE-aware Quantization for LLM Inference Acceleration☆20Jun 27, 2025Updated 8 months ago
- AFPQ code implementation☆23Nov 6, 2023Updated 2 years ago
- FPGA-based hardware acceleration for dropout-based Bayesian Neural Networks.☆27Aug 15, 2023Updated 2 years ago