awilliamson10 / cliporaLinks
Clipora is a powerful toolkit for fine-tuning OpenCLIP models using Low Rank Adapters (LoRA).
☆24Updated last year
Alternatives and similar repositories for clipora
Users that are interested in clipora are comparing it to the libraries listed below
Sorting:
- Code and Models for "GeneCIS A Benchmark for General Conditional Image Similarity"☆60Updated 2 years ago
- Reproducible scaling laws for contrastive language-image learning (https://arxiv.org/abs/2212.07143)☆174Updated 2 months ago
- Code implementation of our NeurIPS 2023 paper: Vocabulary-free Image Classification☆107Updated last year
- An Enhanced CLIP Framework for Learning with Synthetic Captions☆37Updated 4 months ago
- ☆42Updated 7 months ago
- ☆192Updated 2 years ago
- [ICCV 2023] ALIP: Adaptive Language-Image Pre-training with Synthetic Caption☆98Updated last year
- Task Residual for Tuning Vision-Language Models (CVPR 2023)☆74Updated 2 years ago
- [NeurIPS 2023] Text data, code and pre-trained models for paper "Improving CLIP Training with Language Rewrites"☆285Updated last year
- Code and results accompanying our paper titled CHiLS: Zero-Shot Image Classification with Hierarchical Label Sets☆57Updated 2 years ago
- Filtering, Distillation, and Hard Negatives for Vision-Language Pre-Training☆138Updated 2 years ago
- repo for paper titled: Towards Realistic Zero-Shot Classification via Self Structural Semantic Alignment (AAAI'24 Oral)☆25Updated last year
- ☆52Updated 3 years ago
- [ICLR 23] Contrastive Aligned of Vision to Language Through Parameter-Efficient Transfer Learning☆40Updated 2 years ago
- Plotting heatmaps with the self-attention of the [CLS] tokens in the last layer.☆45Updated 3 years ago
- ☆34Updated last year
- Official Code of ECCV 2022 paper MS-CLIP☆90Updated 3 years ago
- Augmenting with Language-guided Image Augmentation (ALIA)☆78Updated last year
- Official repository for "Video-FocalNets: Spatio-Temporal Focal Modulation for Video Action Recognition" [ICCV 2023]☆101Updated last year
- TensorFlow implementation of "TokenLearner: What Can 8 Learned Tokens Do for Images and Videos?"☆35Updated 3 years ago
- ☆120Updated 2 years ago
- COLA: Evaluate how well your vision-language model can Compose Objects Localized with Attributes!☆24Updated 9 months ago
- This is a offical PyTorch/GPU implementation of SupMAE.☆78Updated 3 years ago
- 📍 Official pytorch implementation of paper "ProtoCLIP: Prototypical Contrastive Language Image Pretraining" (IEEE TNNLS)☆53Updated last year
- Code base of SynthCLIP: CLIP training with purely synthetic text-image pairs from LLMs and TTIs.☆100Updated 5 months ago
- Code release for "Understanding Bias in Large-Scale Visual Datasets"☆21Updated 8 months ago
- [ICLR 2024] Official code for the paper "LLM Blueprint: Enabling Text-to-Image Generation with Complex and Detailed Prompts"☆81Updated last year
- ☆30Updated 2 years ago
- Generating Image Specific Text☆28Updated 2 years ago
- ☆11Updated 3 years ago