rajeevsrao / TensorRTLinks
TensorRT is a C++ library for high performance inference on NVIDIA GPUs and deep learning accelerators.
☆19Updated last year
Alternatives and similar repositories for TensorRT
Users that are interested in TensorRT are comparing it to the libraries listed below
Sorting:
- stable diffusion, controlnet, tensorrt, accelerate☆58Updated 2 years ago
- https://wavespeed.ai/ Context parallel attention that accelerates DiT model inference with dynamic caching☆350Updated last month
- Faster generation with text-to-image diffusion models.☆224Updated last month
- End-to-end recipes for optimizing diffusion models with torchao and diffusers (inference and FP8 training).☆372Updated 2 months ago
- Flux diffusion model implementation using quantized fp8 matmul & remaining layers use faster half precision accumulate, which is ~2x fast…☆273Updated 9 months ago
- Experimental usage of stable-fast and TensorRT.☆208Updated last year
- Generate long weighted prompt embeddings for Stable Diffusion☆132Updated 3 months ago
- ☆436Updated last year
- SSD-1B, an open-source text-to-image model, outperforming previous versions by being 50% smaller and 60% faster than SDXL.☆177Updated last year
- Official Repository of the paper "Trajectory Consistency Distillation"☆349Updated last year
- InstantID-ROME: Improved Identity-Preserving Generation in Seconds 🔥☆231Updated last year
- ☆316Updated 2 years ago
- Quantized stable-diffusion cutting down memory 75%, testing in streamlit, deploying in container☆53Updated 2 weeks ago
- ☆285Updated 7 months ago
- [NeurIPS 2024] AsyncDiff: Parallelizing Diffusion Models by Asynchronous Denoising☆202Updated 5 months ago
- Model Compression Toolbox for Large Language Models and Diffusion Models☆578Updated 4 months ago
- Accelerates Flux.1 image generation, just by using this node.☆137Updated 7 months ago
- A gradio web UI demo for Stable Diffusion XL 1.0, with refiner and MultiGPU support☆280Updated last year
- Deploy stable diffusion model with onnx/tenorrt + tritonserver☆124Updated last year
- CPU version of InstantID☆58Updated last year
- ☆55Updated last year
- implementation of the IPAdapter models for HF Diffusers☆175Updated last year
- https://wavespeed.ai/ Best inference performance optimization framework for HuggingFace Diffusers on NVIDIA GPUs.☆1,281Updated 4 months ago
- Segmind Distilled diffusion☆607Updated last year
- ☆139Updated last week
- ☆50Updated last year
- Real-time inference for Stable Diffusion - 0.88s latency. Covers AITemplate, nvFuser, TensorRT, FlashAttention. Join our Discord communty…☆558Updated last year
- [CVPR 2024] DeepCache: Accelerating Diffusion Models for Free☆914Updated last year
- Tiny optimized Stable-diffusion that can run on GPUs with just 1GB of VRAM. (Beta)☆174Updated 2 years ago
- ☆119Updated last year