cardinalblue / clip-models-for-distillationLinks
☆18Updated last year
Alternatives and similar repositories for clip-models-for-distillation
Users that are interested in clip-models-for-distillation are comparing it to the libraries listed below
Sorting:
- ☆26Updated 4 years ago
- source code and pre-trained/fine-tuned checkpoint for NAACL 2021 paper LightningDOT☆72Updated 2 years ago
- [FGVC9-CVPR 2022] The second place solution for 2nd eBay eProduct Visual Search Challenge.☆26Updated 2 years ago
- A non-JIT version implementation / replication of CLIP of OpenAI in pytorch☆34Updated 4 years ago
- ☆11Updated 4 years ago
- A unified framework to jointly model images, text, and human attention traces.☆78Updated 4 years ago
- Research code for "Training Vision-Language Transformers from Captions Alone"☆34Updated 2 years ago
- MDMMT: Multidomain Multimodal Transformer for Video Retrieval☆26Updated 3 years ago
- ECCV2020 paper: Fashion Captioning: Towards Generating Accurate Descriptions with Semantic Rewards. Code and Data.☆85Updated 2 years ago
- ☆32Updated 3 years ago
- ☆47Updated last month
- ☆20Updated 3 years ago
- Repository for the paper "Data Efficient Masked Language Modeling for Vision and Language".☆18Updated 3 years ago
- This project provides a data set with bounding boxes, body poses, 3D face meshes & captions of people from our LAION-2.2B. Additionally i…☆14Updated 3 years ago
- MLPs for Vision and Langauge Modeling (Coming Soon)☆27Updated 3 years ago
- Use CLIP to represent video for Retrieval Task☆69Updated 4 years ago
- ROSITA: Enhancing Vision-and-Language Semantic Alignments via Cross- and Intra-modal Knowledge Integration☆56Updated 2 years ago
- This is the official repository for CookGAN: Meal Image Synthesis from Ingredients☆23Updated 2 years ago
- Implementation of our PR 2020 paper:Unsupervised Text-to-Image Synthesis☆13Updated 4 years ago
- [ICME 2022] code for the paper, SimVit: Exploring a simple vision transformer with sliding windows.☆68Updated 2 years ago
- ☆9Updated 2 years ago
- This dataset contains about 110k images annotated with the depth and occlusion relationships between arbitrary objects. It enables resear…☆16Updated 4 years ago
- [EMNLP 2021] Code and data for our paper "Vision-and-Language or Vision-for-Language? On Cross-Modal Influence in Multimodal Transformers…☆20Updated 3 years ago
- ☆46Updated 3 years ago
- ☆28Updated 5 years ago
- ☆28Updated 3 years ago
- Vision Longformer For Object Detection☆34Updated 4 years ago
- [CVPR 2022 Challenge Rank 1st] The official code for V2L: Leveraging Vision and Vision-language Models into Large-scale Product Retrieval…☆29Updated 2 years ago
- codebase for the SIMAT dataset and evaluation☆39Updated 3 years ago
- CLIP-Art: Contrastive Pre-training for Fine-Grained Art Classification - 4th Workshop on Computer Vision for Fashion, Art, and Design☆27Updated 3 years ago