turingmotors / swan
This project aims to enable language model inference on FPGAs, supporting AI applications in edge devices and environments with limited resources.
☆137Updated 6 months ago
Related projects ⓘ
Alternatives and complementary repositories for swan
- Deep learning accelerator architectures requiring half the multipliers☆263Updated 7 months ago
- Run 64-bit Linux on LiteX + RocketChip☆188Updated 3 months ago
- Sequential Logic☆99Updated this week
- A configurable RTL to bitstream FPGA toolchain☆256Updated this week
- Code sample showing how to run and benchmark models on Qualcomm's Window PCs☆87Updated last month
- Hashed Lookup Table based Matrix Multiplication (halutmatmul) - Stella Nera accelerator☆207Updated 11 months ago
- Repository for the QUIK project, enabling the use of 4bit kernels for generative inference - EMNLP 2024☆173Updated 7 months ago
- Exploring the scalable matrix extension of the Apple M4 processor☆135Updated 2 weeks ago
- Open source machine learning accelerators☆360Updated 8 months ago
- Veryl: A Modern Hardware Description Language☆513Updated this week
- minimal yet working VPN daemon for Linux☆107Updated last week
- ☆74Updated 10 months ago
- Tiny ASIC implementation for "The Era of 1-bit LLMs All Large Language Models are in 1.58 Bits" matrix multiplication unit☆111Updated 7 months ago
- Tenstorrent console based hardware information program☆23Updated 2 weeks ago
- The Riallto Open Source Project from AMD☆69Updated last week
- An implementation of bucketMul LLM inference☆214Updated 4 months ago
- Fork of LLVM to support AMD AIEngine processors☆107Updated this week
- Universal Memory Interface (UMI)☆142Updated last week
- Repository of model demos using TT-Buda☆55Updated 3 weeks ago
- 🪝"mnist" in 60 lines of code, no dependencies. For educational purposes.☆31Updated 4 months ago
- +256,000,000 points per plot, +60 Fps on shity laptop. Only limit is the size of your RAM.☆144Updated last week
- Absolute minimalistic implementation of a GPT-like transformer using only numpy (<650 lines).☆250Updated last year
- Binary Neural Network Framework for FPGA(Differentiable LUT)☆139Updated last week
- A CLI to manage install and configure llama inference implemenation in multiple languages☆65Updated 10 months ago
- ☆234Updated 8 months ago
- NNgen: A Fully-Customizable Hardware Synthesis Compiler for Deep Neural Network☆339Updated last year
- Run and explore Llama models locally with minimal dependencies on CPU☆183Updated last month
- ☆179Updated 2 months ago
- ☆224Updated last month
- Transformer GPU VRAM estimator☆40Updated 7 months ago