haltakov / natural-language-image-search
Search photos on Unsplash using natural language
☆987Updated 2 years ago
Related projects ⓘ
Alternatives and complementary repositories for natural-language-image-search
- OpenAI CLIP text encoders for multiple languages!☆763Updated last year
- Open-AI's DALL-E for large scale training in mesh-tensorflow.☆434Updated 2 years ago
- Crop using CLIP☆336Updated 2 years ago
- Simple image captioning model☆1,317Updated 5 months ago
- Easily compute clip embeddings and build a clip retrieval system with them☆2,413Updated 7 months ago
- WIT (Wikipedia-based Image Text) Dataset is a large multimodal multilingual dataset comprising 37M+ image-text sets with 11M+ unique imag…☆1,014Updated last month
- Easily turn large sets of image urls to an image dataset. Can download, resize and package 100M urls in 20h on one machine.☆3,723Updated 3 months ago
- Search photos on Unsplash based on OpenAI's CLIP model, support search with joint image+text queries and attention visualization.☆209Updated 3 years ago
- Search images with a text or image query, using Open AI's pretrained CLIP model.☆205Updated 2 years ago
- Contrastive Language-Image Forensic Search allows free text searching through videos using OpenAI's machine learning model CLIP☆451Updated 2 years ago
- Official Implementation for "StyleCLIP: Text-Driven Manipulation of StyleGAN Imagery" (ICCV 2021 Oral)☆4,004Updated last year
- Officially unofficial re-implementation of paper: Paint Transformer: Feed Forward Neural Painting with Stroke Prediction, ICCV 2021.☆490Updated last year
- Code release for SLIP Self-supervision meets Language-Image Pre-training☆747Updated last year
- Text-to-Image generation. The repo for NeurIPS 2021 paper "CogView: Mastering Text-to-Image Generation via Transformers".☆1,724Updated last year
- Official repository of OFA (ICML 2022). Paper: OFA: Unifying Architectures, Tasks, and Modalities Through a Simple Sequence-to-Sequence L…☆2,421Updated 6 months ago
- Conceptual 12M is a dataset containing (image-URL, caption) pairs collected for vision-and-language pre-training.☆368Updated last year
- official code repo for paper "CogView2: Faster and Better Text-to-Image Generation via Hierarchical Transformers"☆951Updated 2 years ago
- Implementation / replication of DALL-E, OpenAI's Text to Image Transformer, in Pytorch☆5,572Updated 9 months ago
- Oscar and VinVL☆1,038Updated last year
- Awesome list for research on CLIP (Contrastive Language-Image Pre-Training).☆1,136Updated 4 months ago
- A PyTorch Lightning solution to training OpenAI's CLIP from scratch.☆665Updated 2 years ago
- [CVPR 2023] Official Implementation of X-Decoder for generalized decoding for pixel, image and language☆1,289Updated last year
- PyTorch code for "Fine-grained Image Captioning with CLIP Reward" (Findings of NAACL 2022)☆235Updated 2 years ago
- ☆1,157Updated 2 years ago
- Implementation of CoCa, Contrastive Captioners are Image-Text Foundation Models, in Pytorch☆1,067Updated 11 months ago
- A concise but complete implementation of CLIP with various experimental improvements from recent papers☆693Updated last year
- GIT: A Generative Image-to-text Transformer for Vision and Language☆549Updated 11 months ago
- CLIP (Contrastive Language–Image Pre-training) for Italian☆181Updated last year
- Implementation of 🦩 Flamingo, state-of-the-art few-shot visual question answering attention net out of Deepmind, in Pytorch☆1,215Updated 2 years ago
- COYO-700M: Large-scale Image-Text Pair Dataset☆1,163Updated last year