leejet / stable-diffusion.cpp
Stable Diffusion and Flux in pure C/C++
☆4,064Updated 2 months ago
Alternatives and similar repositories for stable-diffusion.cpp
Users that are interested in stable-diffusion.cpp are comparing it to the libraries listed below
Sorting:
- Tensor library for machine learning☆12,477Updated this week
- INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model☆1,514Updated last month
- Run GGUF models easily with a KoboldAI UI. One File. Zero Install.☆7,246Updated this week
- A fast inference library for running LLMs locally on modern consumer-class GPUs☆4,166Updated this week
- Lightweight inference library for ONNX files, written in C++. It can run Stable Diffusion XL 1.0 on a RPI Zero 2 (or in 298MB of RAM) but…☆1,940Updated 2 weeks ago
- Fast stable diffusion on CPU☆1,684Updated last week
- This repository contains a pure C++ ONNX implementation of multiple offline AI models, such as StableDiffusion (1.5 and XL), ControlNet, …☆614Updated last year
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆2,871Updated last year
- ☆1,275Updated last year
- Python bindings for llama.cpp☆9,063Updated last month
- Suno AI's Bark model in C/C++ for fast text-to-speech generation☆809Updated 5 months ago
- lightweight, standalone C++ inference engine for Google's Gemma models.☆6,396Updated this week
- RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable)…☆13,576Updated this week
- Simple, safe way to store and distribute tensors☆3,257Updated this week
- Inference Llama 2 in one file of pure C☆18,371Updated 9 months ago
- High-speed Large Language Model Serving for Local Deployment☆8,191Updated 2 months ago
- CLIP inference in plain C/C++ with no extra dependencies☆498Updated 8 months ago
- The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.☆8,473Updated last year
- Llama 2 Everywhere (L2E)☆1,517Updated 3 months ago
- Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)☆568Updated last year
- Python bindings for the Transformer models implemented in C/C++ using GGML library.☆1,862Updated last year
- LLM inference in C/C++☆79,738Updated this week
- An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.☆4,838Updated last month
- Fast inference engine for Transformer models☆3,788Updated last month
- ☆7,805Updated last year
- https://wavespeed.ai/ Best inference performance optimization framework for HuggingFace Diffusers on NVIDIA GPUs.☆1,254Updated last month
- StableSwarmUI, A Modular Stable Diffusion Web-User-Interface, with an emphasis on making powertools easily accessible, high performance, …☆4,783Updated 2 months ago
- 4 bits quantization of LLaMA using GPTQ☆3,050Updated 10 months ago
- Port of OpenAI's Whisper model in C/C++☆39,829Updated this week
- Bringing stable diffusion models to web browsers. Everything runs inside the browser with no server support.☆3,657Updated last year