openai / CLIP-featurevisLinks
code for reproducing some of the diagrams in the paper "Multimodal Neurons in Artificial Neural Networks"
☆308Updated 4 years ago
Alternatives and similar repositories for CLIP-featurevis
Users that are interested in CLIP-featurevis are comparing it to the libraries listed below
Sorting:
- ☆234Updated 2 years ago
- Repository for "Generating images from caption and vice versa via CLIP-Guided Generative Latent Space Search"☆178Updated 3 years ago
- Code for making music videos using CLIP☆174Updated 4 years ago
- Official codebase for Pretrained Transformers as Universal Computation Engines.☆249Updated 3 years ago
- Learning to ground explanations of affect for visual art.☆316Updated 4 years ago
- This is a summary of easily available datasets for generalized DALLE-pytorch training.☆128Updated 3 years ago
- Open-AI's DALL-E for large scale training in mesh-tensorflow.☆434Updated 3 years ago
- PyTorch Implementation of OpenAI's Image GPT☆258Updated last year
- An attempt at the implementation of Glom, Geoffrey Hinton's new idea that integrates concepts from neural fields, top-down-bottom-up proc…☆193Updated 4 years ago
- ☆87Updated 3 years ago
- ☆351Updated 3 years ago
- Here is a collection of checkpoints for DALLE-pytorch models, from where you can keep on training or start generating images.☆146Updated 2 years ago
- ☆198Updated 3 years ago
- Contrastive Language-Image Pretraining☆143Updated 2 years ago
- Code to train and evaluate the GeNeVA-GAN model for the GeNeVA task proposed in our ICCV 2019 paper "Tell, Draw, and Repeat: Generating a…☆85Updated 2 years ago
- Rewriting a Deep Generative Model, ECCV 2020 (oral). Interactive tool to directly edit the rules of a GAN to synthesize scenes with obje…☆536Updated last year
- Code for the paper, "Distribution Augmentation for Generative Modeling", ICML 2020.☆125Updated 2 years ago
- ImageBART: Bidirectional Context with Multinomial Diffusion for Autoregressive Image Synthesis☆125Updated 3 years ago
- ☆160Updated 3 years ago
- Conceptual 12M is a dataset containing (image-URL, caption) pairs collected for vision-and-language pre-training.☆396Updated 2 years ago
- Modelverse: Content-Based Search for Deep Generative Models☆224Updated 7 months ago
- ☆152Updated last year
- Network-to-Network Translation with Conditional Invertible Neural Networks☆225Updated 2 years ago
- Multi-object image datasets with ground-truth segmentation masks and generative factors.☆272Updated 3 years ago
- Implementation of NÜWA, state of the art attention network for text to video synthesis, in Pytorch☆550Updated 2 years ago
- [CVPR 2021] VirTex: Learning Visual Representations from Textual Annotations☆563Updated last year
- Trains Transformer model variants. Data isn't shuffled between batches.☆143Updated 2 years ago
- Refactoring dalle-pytorch and taming-transformers for TPU VM☆60Updated 3 years ago
- MERLOT: Multimodal Neural Script Knowledge Models☆224Updated 3 years ago
- JAX implementation of VQGAN☆93Updated 3 years ago