leejet / stable-diffusion.cppLinks
Diffusion model(SD,Flux,Wan,Qwen Image,Z-Image,...) inference in pure C/C++
☆4,899Updated this week
Alternatives and similar repositories for stable-diffusion.cpp
Users that are interested in stable-diffusion.cpp are comparing it to the libraries listed below
Sorting:
- Tensor library for machine learning☆13,714Updated this week
- Fast stable diffusion on CPU and AI PC☆1,913Updated 3 weeks ago
- A fast inference library for running LLMs locally on modern consumer-class GPUs☆4,388Updated last week
- Lightweight inference library for ONNX files, written in C++. It can run Stable Diffusion XL 1.0 on a RPI Zero 2 (or in 298MB of RAM) but…☆2,011Updated last month
- Python bindings for llama.cpp☆9,821Updated 4 months ago
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆2,904Updated 2 years ago
- INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model☆1,557Updated 8 months ago
- Local AI API Platform☆2,764Updated 5 months ago
- High-speed Large Language Model Serving for Local Deployment☆8,460Updated 4 months ago
- llama.cpp fork with additional SOTA quants and improved performance☆1,390Updated this week
- Suno AI's Bark model in C/C++ for fast text-to-speech generation☆848Updated last year
- This repository contains a pure C++ ONNX implementation of multiple offline AI models, such as StableDiffusion (1.5 and XL), ControlNet, …☆629Updated 6 months ago
- ☆1,278Updated 2 years ago
- Simple, safe way to store and distribute tensors☆3,557Updated this week
- Pure C++ implementation of several models for real-time chatting on your computer (CPU & GPU)☆757Updated this week
- Official Code for Stable Cascade☆6,588Updated last year
- LLM inference in C/C++☆91,520Updated this week
- Python bindings for the Transformer models implemented in C/C++ using GGML library.☆1,877Updated last year
- Large-scale LLM inference engine☆1,607Updated 3 weeks ago
- AMD-SHARK Studio -- Web UI for SHARK+IREE High Performance Machine Learning Distribution☆1,451Updated this week
- Run GGUF models easily with a KoboldAI UI. One File. Zero Install.☆9,080Updated this week
- https://wavespeed.ai/ Best inference performance optimization framework for HuggingFace Diffusers on NVIDIA GPUs.☆1,296Updated 8 months ago
- lightweight, standalone C++ inference engine for Google's Gemma models.☆6,643Updated this week
- 🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading☆9,858Updated last year
- Distributed LLM inference. Connect home devices into a powerful cluster to accelerate LLM inference. More devices means faster inference.☆2,770Updated 2 weeks ago
- Blazingly fast LLM inference.☆6,280Updated this week
- CLIP inference in plain C/C++ with no extra dependencies☆543Updated 5 months ago
- AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (N…☆4,697Updated last month
- Inference Llama 2 in one file of pure C☆19,032Updated last year
- Stable Diffusion in NCNN with c++, supported txt2img and img2img☆1,054Updated 2 years ago