vitoplantamura / OnnxStreamLinks
Lightweight inference library for ONNX files, written in C++. It can run Stable Diffusion XL 1.0 on a RPI Zero 2 (or in 298MB of RAM) but also Mistral 7B on desktops and servers. ARM, x86, WASM, RISC-V supported. Accelerated by XNNPACK.
☆1,972Updated 2 weeks ago
Alternatives and similar repositories for OnnxStream
Users that are interested in OnnxStream are comparing it to the libraries listed below
Sorting:
- Llama 2 Everywhere (L2E)☆1,519Updated 6 months ago
- Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)☆568Updated 2 years ago
- ☆1,276Updated last year
- This repository contains a pure C++ ONNX implementation of multiple offline AI models, such as StableDiffusion (1.5 and XL), ControlNet, …☆622Updated 2 months ago
- Stable Diffusion and Flux in pure C/C++☆4,293Updated last week
- Fork of Facebooks LLaMa model to run on CPU☆772Updated 2 years ago
- ☆1,028Updated last year
- C++ implementation for BLOOM☆810Updated 2 years ago
- INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model☆1,537Updated 4 months ago
- An extensible, easy-to-use, and portable diffusion web UI 👨🎨☆1,674Updated last year
- Fast stable diffusion on CPU and AI PC☆1,756Updated last week
- CLIP inference in plain C/C++ with no extra dependencies☆515Updated last month
- MiniLLM is a minimal system for running modern LLMs on consumer-grade GPUs☆918Updated 2 years ago
- Bringing stable diffusion models to web browsers. Everything runs inside the browser with no server support.☆3,677Updated last year
- Connect home devices into a powerful cluster to accelerate LLM inference. More devices means faster inference.☆2,235Updated this week
- Suno AI's Bark model in C/C++ for fast text-to-speech generation☆836Updated 8 months ago
- Raspberry Pi Voice Assistant☆796Updated 7 months ago
- Simple UI for LLM Model Finetuning☆2,061Updated last year
- A fast inference library for running LLMs locally on modern consumer-class GPUs☆4,258Updated 3 weeks ago
- Use Code Llama with Visual Studio Code and the Continue extension. A local LLM alternative to GitHub Copilot.☆569Updated last year
- Quantized inference code for LLaMA models☆1,049Updated 2 years ago
- Local AI API Platform☆2,765Updated last month
- A diffusion model to colorize black and white images☆779Updated 2 years ago
- A "large" language model running on a microcontroller☆533Updated last year
- Instruct-tune LLaMA on consumer hardware☆362Updated 2 years ago
- A simple "Be My Eyes" web app with a llama.cpp/llava backend☆490Updated last year
- Cross-Platform, GPU Accelerated Whisper 🏎️☆1,802Updated last year
- ☆1,540Updated last year
- llama3.np is a pure NumPy implementation for Llama 3 model.☆987Updated 3 months ago
- Explore large language models in 512MB of RAM☆1,197Updated 2 weeks ago