kamalkraj / stable-diffusion-tritonserver
Deploy stable diffusion model with onnx/tenorrt + tritonserver
☆123Updated last year
Alternatives and similar repositories for stable-diffusion-tritonserver:
Users that are interested in stable-diffusion-tritonserver are comparing it to the libraries listed below
- ☆54Updated 2 years ago
- The Triton backend for TensorRT.☆73Updated this week
- Faster generation with text-to-image diffusion models.☆213Updated 6 months ago
- SSD-1B, an open-source text-to-image model, outperforming previous versions by being 50% smaller and 60% faster than SDXL.☆175Updated last year
- NVIDIA® TensorRT™, an SDK for high-performance deep learning inference, includes a deep learning inference optimizer and runtime that del…☆26Updated last year
- An efficient implementation of Stable-Diffusion-XL☆46Updated last year
- stable diffusion, controlnet, tensorrt, accelerate☆56Updated last year
- Generate long weighted prompt embeddings for Stable Diffusion☆112Updated 7 months ago
- The Triton backend for the ONNX Runtime.☆140Updated last week
- ☆99Updated last year
- Unofficial implementation. Stable diffusion model trained by AI Feedback-Based Self-Training Direct Preference Optimization.☆63Updated last year
- A Gradio component that can be used to annotate images with bounding boxes.☆49Updated last month
- 🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch☆50Updated 2 years ago
- Simple large-scale training of stable diffusion with multi-node support.☆131Updated last year
- Real-time inference for Stable Diffusion - 0.88s latency. Covers AITemplate, nvFuser, TensorRT, FlashAttention. Join our Discord communty…☆557Updated last year
- Huggingface-compatible SDXL Unet implementation that is readily hackable☆416Updated last year
- https://wavespeed.ai/ Context parallel attention that accelerates DiT model inference with dynamic caching☆243Updated 3 weeks ago
- It is a simple library to speed up CLIP inference up to 3x (K80 GPU)☆217Updated last year
- TensorRT is a C++ library for high performance inference on NVIDIA GPUs and deep learning accelerators.☆19Updated last year
- [WIP] Better (FP8) attention for Hopper☆30Updated 2 months ago
- Triton Model Navigator is an inference toolkit designed for optimizing and deploying Deep Learning models with a focus on NVIDIA GPUs.☆199Updated 3 months ago
- ☆246Updated last week
- Iterable datapipelines for pytorch training.☆83Updated 7 months ago
- End-to-end recipes for optimizing diffusion models with torchao and diffusers (inference and FP8 training).☆342Updated 2 months ago
- Quantized stable-diffusion cutting down memory 75%, testing in streamlit, deploying in container☆54Updated 2 weeks ago
- Diffusers training with mmengine☆100Updated last year
- faster parallel inference of mochi-1 video generation model☆117Updated 2 months ago
- Official Pytorch implementation of "Graphit: A Unified Framework for Diverse Image Editing Tasks"☆201Updated last year
- OpenAI compatible API for TensorRT LLM triton backend☆205Updated 8 months ago
- Common source, scripts and utilities shared across all Triton repositories.☆69Updated last week