timudk / flux_triton
Writing FLUX in Triton
☆32Updated 6 months ago
Alternatives and similar repositories for flux_triton:
Users that are interested in flux_triton are comparing it to the libraries listed below
- Minimal Differentiable Image Reward Functions☆53Updated this week
- ☆32Updated 5 months ago
- ☆22Updated 9 months ago
- ☆28Updated 8 months ago
- (WIP) Parallel inference for black-forest-labs' FLUX model.☆18Updated 4 months ago
- WIP Pytorch code for stably training single-step, mode-dropping, deterministic autoencoders☆25Updated 11 months ago
- [WIP] Better (FP8) attention for Hopper☆27Updated last month
- Official repository for VQDM:Accurate Compression of Text-to-Image Diffusion Models via Vector Quantization paper☆33Updated 6 months ago
- ☆25Updated 10 months ago
- Triton kernels for Flux☆20Updated 3 months ago
- Official codebase for Margin-aware Preference Optimization for Aligning Diffusion Models without Reference (MaPO).☆72Updated 10 months ago
- ☆27Updated 11 months ago
- ☆36Updated 11 months ago
- Focused on fast experimentation and simplicity☆71Updated 3 months ago
- Implementation of SmoothCache, a project aimed at speeding-up Diffusion Transformer (DiT) based GenAI models with error-guided caching.☆42Updated 3 weeks ago
- ☆49Updated last year
- ☆65Updated 3 months ago
- Fine-tune of Florence-2 for shot categorization.☆24Updated last month
- Recaption large (Web)Datasets with vllm and save the artifacts.☆50Updated 4 months ago
- PyTorch half precision gemm lib w/ fused optional bias + optional relu/gelu☆61Updated 4 months ago
- research impl of Native Sparse Attention (2502.11089)☆53Updated last month
- ☆13Updated 10 months ago
- This repository shows how to use Q8 kernels with `diffusers` to optimize inference of LTX-Video on ADA GPUs.☆17Updated 3 months ago
- Train transformer language models with reinforcement learning.☆18Updated last year
- Distilling Diversity and Control in Diffusion Models☆37Updated 2 weeks ago
- LoRA fine-tune directly on the quantized models.☆27Updated 4 months ago
- A comprehensive codebase for training and finetuning Image <> Latent models.☆31Updated last month
- TerDiT: Ternary Diffusion Models with Transformers☆70Updated 9 months ago
- Omegance: A Single Parameter for Various Granularities in Diffusion-Based Synthesis (arXiv, 2024)☆50Updated 4 months ago
- Implementation of the proposed MaskBit from Bytedance AI☆75Updated 5 months ago