rejunity / tiny-asic-1_58bit-matrix-mulLinks
Tiny ASIC implementation for "The Era of 1-bit LLMs All Large Language Models are in 1.58 Bits" matrix multiplication unit
☆166Updated last year
Alternatives and similar repositories for tiny-asic-1_58bit-matrix-mul
Users that are interested in tiny-asic-1_58bit-matrix-mul are comparing it to the libraries listed below
Sorting:
- ☆108Updated last year
- The Riallto Open Source Project from AMD☆84Updated 6 months ago
- Machine-Learning Accelerator System Exploration Tools☆179Updated 3 weeks ago
- An AI accelerator implementation with Xilinx FPGA☆64Updated 9 months ago
- ShiftAddLLM: Accelerating Pretrained LLMs via Post-Training Multiplication-Less Reparameterization☆110Updated last year
- A high-efficiency system-on-chip for floating-point compute workloads.☆43Updated 9 months ago
- ☆33Updated last year
- A new LLM solution for RTL code generation, achieving state-of-the-art performance in non-commercial solutions and outperforming GPT-3.5.☆232Updated 8 months ago
- A survey on Hardware Accelerated LLMs☆59Updated 9 months ago
- A minimal Tensor Processing Unit (TPU) inspired by Google's TPUv1.☆186Updated last year
- DNN Compiler for Heterogeneous SoCs☆52Updated last week
- Verilog evaluation benchmark for large language model☆331Updated 3 months ago
- Torch2Chip (MLSys, 2024)☆54Updated 7 months ago
- Ocelot: The Berkeley Out-of-Order Machine With V-EXT support☆189Updated last week
- Research and Materials on Hardware implementation of Transformer Model☆285Updated 8 months ago
- Floating point modules for CHISEL☆31Updated 11 years ago
- ☆38Updated 7 months ago
- NeuraLUT-Assemble☆43Updated 2 months ago
- Verilog package manager written in Rust☆143Updated last year
- PDPU: An Open-Source Posit Dot-Product Unit for Deep Learning Applications☆43Updated 2 years ago
- Attention in SRAM on Tenstorrent Grayskull☆38Updated last year
- ☆49Updated last month
- MAGE: A Multi-Agent Engine for Automated RTL Code Generation☆67Updated 6 months ago
- Fully opensource spiking neural network accelerator☆158Updated 2 years ago
- Prepare for DeekSeek R1 inference: Benchmark CPU, DRAM, SSD, iGPU, GPU, ... with efficient code.☆74Updated 9 months ago
- An Open Workflow to Build Custom SoCs and run Deep Models at the Edge☆96Updated this week
- Universal Memory Interface (UMI)☆153Updated last week
- Open source machine learning accelerators☆387Updated last year
- LLM Agent for Hardware Description Language☆20Updated 4 months ago
- Fork of upstream onnxruntime focused on supporting risc-v accelerators☆87Updated 2 years ago