leejet / stable-diffusion.cppLinks
Diffusion model(SD,Flux,Wan,Qwen Image,...) inference in pure C/C++
☆4,453Updated this week
Alternatives and similar repositories for stable-diffusion.cpp
Users that are interested in stable-diffusion.cpp are comparing it to the libraries listed below
Sorting:
- Tensor library for machine learning☆13,261Updated last week
- Lightweight inference library for ONNX files, written in C++. It can run Stable Diffusion XL 1.0 on a RPI Zero 2 (or in 298MB of RAM) but…☆1,992Updated 2 weeks ago
- Fast stable diffusion on CPU and AI PC☆1,803Updated last month
- A fast inference library for running LLMs locally on modern consumer-class GPUs☆4,341Updated 2 months ago
- INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model☆1,547Updated 6 months ago
- Suno AI's Bark model in C/C++ for fast text-to-speech generation☆839Updated 11 months ago
- Llama 2 Everywhere (L2E)☆1,522Updated last month
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆2,903Updated 2 years ago
- ☆1,277Updated last year
- Stable Diffusion in NCNN with c++, supported txt2img and img2img☆1,049Updated 2 years ago
- This repository contains a pure C++ ONNX implementation of multiple offline AI models, such as StableDiffusion (1.5 and XL), ControlNet, …☆624Updated 4 months ago
- Python bindings for llama.cpp☆9,658Updated 2 months ago
- High-speed Large Language Model Serving for Local Deployment☆8,367Updated 2 months ago
- Local AI API Platform☆2,760Updated 3 months ago
- StableSwarmUI, A Modular Stable Diffusion Web-User-Interface, with an emphasis on making powertools easily accessible, high performance, …☆4,866Updated 7 months ago
- https://wavespeed.ai/ Best inference performance optimization framework for HuggingFace Diffusers on NVIDIA GPUs.☆1,287Updated 6 months ago
- Pure C++ implementation of several models for real-time chatting on your computer (CPU & GPU)☆720Updated this week
- Inference Llama 2 in one file of pure C☆18,848Updated last year
- CLIP inference in plain C/C++ with no extra dependencies☆523Updated 4 months ago
- SHARK Studio -- Web UI for SHARK+IREE High Performance Machine Learning Distribution☆1,449Updated 6 months ago
- SD.Next: All-in-one WebUI for AI generative image and video creation☆6,648Updated last week
- Python bindings for the Transformer models implemented in C/C++ using GGML library.☆1,876Updated last year
- Run any Llama 2 locally with gradio UI on GPU or CPU from anywhere (Linux/Windows/Mac). Use `llama2-wrapper` as your local llama2 backend…☆1,956Updated last year
- Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)☆568Updated 2 years ago
- The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.☆8,768Updated last year
- Bringing stable diffusion models to web browsers. Everything runs inside the browser with no server support.☆3,685Updated last year
- Official Code for Stable Cascade☆6,585Updated last year
- Large-scale LLM inference engine☆1,562Updated last week
- An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.☆4,965Updated 6 months ago
- llama.cpp fork with additional SOTA quants and improved performance☆1,258Updated this week