kamalkraj / stable-diffusion-tritonserver
Deploy stable diffusion model with onnx/tenorrt + tritonserver
☆123Updated last year
Alternatives and similar repositories for stable-diffusion-tritonserver:
Users that are interested in stable-diffusion-tritonserver are comparing it to the libraries listed below
- ☆52Updated last year
- The Triton backend for TensorRT.☆68Updated this week
- Faster generation with text-to-image diffusion models.☆206Updated 3 months ago
- stable diffusion, controlnet, tensorrt, accelerate☆55Updated last year
- This is a Gradio WebUI working with the Diffusers format of Stable Diffusion☆79Updated 2 years ago
- It is a simple library to speed up CLIP inference up to 3x (K80 GPU)☆209Updated last year
- SSD-1B, an open-source text-to-image model, outperforming previous versions by being 50% smaller and 60% faster than SDXL.☆168Updated 9 months ago
- TensorRT is a C++ library for high performance inference on NVIDIA GPUs and deep learning accelerators.☆17Updated 10 months ago
- NVIDIA® TensorRT™, an SDK for high-performance deep learning inference, includes a deep learning inference optimizer and runtime that del…☆25Updated last year
- ☆98Updated last year
- Python bindings for ggml☆136Updated 4 months ago
- Generate long weighted prompt embeddings for Stable Diffusion☆101Updated 3 months ago
- Quantized stable-diffusion cutting down memory 75%, testing in streamlit, deploying in container☆54Updated this week
- ☆30Updated 2 years ago
- ☆107Updated 2 years ago
- A repository containing datasets and tools to train a watermark classifier.☆63Updated 2 years ago
- Common source, scripts and utilities shared across all Triton repositories.☆66Updated this week
- End-to-end recipes for optimizing diffusion models with torchao and diffusers (inference and FP8 training).☆307Updated 2 weeks ago
- The Triton backend for the ONNX Runtime.☆136Updated this week
- PyTriton is a Flask/FastAPI-like interface that simplifies Triton's deployment in Python environments.☆763Updated 2 months ago
- Iterable datapipelines for pytorch training.☆81Updated 4 months ago
- Triton Model Navigator is an inference toolkit designed for optimizing and deploying Deep Learning models with a focus on NVIDIA GPUs.☆193Updated this week
- OpenAI compatible API for TensorRT LLM triton backend☆186Updated 5 months ago
- Model Compression Toolbox for Large Language Models and Diffusion Models☆302Updated 3 weeks ago
- Real-time inference for Stable Diffusion - 0.88s latency. Covers AITemplate, nvFuser, TensorRT, FlashAttention. Join our Discord communty…☆554Updated last year
- Simple example of FastAPI + gRPC AsyncIO + Triton☆61Updated 2 years ago
- Optimum version of a UI for Stable Diffusion, running on ONNX models for faster inference, working on most common GPU vendors: NVIDIA,AMD…☆22Updated last year
- An efficient implementation of Stable-Diffusion-XL☆45Updated last year
- what I learned about fine-tuning stable diffusion☆140Updated last year
- Common source, scripts and utilities for creating Triton backends.☆305Updated this week