shashnkvats / IndofashionclipLinks
Fine tuning OpenAI's CLIP model on Indian Fashion Dataset
☆51Updated 2 years ago
Alternatives and similar repositories for Indofashionclip
Users that are interested in Indofashionclip are comparing it to the libraries listed below
Sorting:
- Finetuning CLIP on a small image/text dataset using huggingface libs☆51Updated 2 years ago
- Image Prompter for Gradio☆92Updated last year
- [ICCV2023] Segment Every Reference Object in Spatial and Temporal Spaces☆238Updated 7 months ago
- [ACM TOMM 2023] - Composed Image Retrieval using Contrastive Learning and Task-oriented CLIP-based Features☆184Updated 2 years ago
- Projects based on SigLIP (Zhai et. al, 2023) and Hugging Face transformers integration 🤗☆275Updated 7 months ago
- Segment Anything combined with CLIP☆347Updated last year
- GroundedSAM Base Model plugin for Autodistill☆52Updated last year
- Fine-tuning OpenAI CLIP Model for Image Search on medical images☆77Updated 3 years ago
- Fine-tuning code for CLIP models☆253Updated 2 months ago
- Data release for the ImageInWords (IIW) paper.☆220Updated 10 months ago
- A simple Segment Anything WebUI based on Gradio.☆81Updated 2 years ago
- Image Editing Anything☆116Updated 2 years ago
- A component that allows you to annotate an image with points and boxes.☆21Updated last year
- Few shot recognition using CLIP's OpenAI architecture.☆36Updated 4 years ago
- Easiest way of fine-tuning HuggingFace video classification models☆145Updated 2 years ago
- The official repo for the paper "VeCLIP: Improving CLIP Training via Visual-enriched Captions"☆246Updated 7 months ago
- Use Grounding DINO, Segment Anything, and GPT-4V to label images with segmentation masks for use in training smaller, fine-tuned models.☆66Updated last year
- CLIPxGPT Captioner is Image Captioning Model based on OpenAI's CLIP and GPT-2.☆118Updated 7 months ago
- Estimate dataset difficulty and detect label mistakes using reconstruction error ratios!☆26Updated 8 months ago
- [CVPR 24] The repository provides code for running inference and training for "Segment and Caption Anything" (SCA) , links for downloadin…☆228Updated 11 months ago
- Implementation of PALI3 from the paper PALI-3 VISION LANGUAGE MODELS: SMALLER, FASTER, STRONGER"☆145Updated 2 weeks ago
- This is a public repository for Image Clustering Conditioned on Text Criteria (IC|TC)☆90Updated last year
- Codebase for the Recognize Anything Model (RAM)☆83Updated last year
- FInetuning CLIP for Few Shot Learning☆45Updated 3 years ago
- State-of-the-art CLIP/SigLIP embedding models finetuned for the fashion domain. +57% increase in evaluation metrics vs FashionCLIP 2.0.☆106Updated last year
- A Simple Image Clustering Script using CLIP and Hierarchial Clustering☆38Updated 2 years ago
- This is an implementation of zero-shot instance segmentation using Segment Anything.☆314Updated 2 years ago
- Holds code for our CVPR'23 tutorial: All Things ViTs: Understanding and Interpreting Attention in Vision.☆195Updated 2 years ago
- [ICLR 2024] Official code for the paper "LLM Blueprint: Enabling Text-to-Image Generation with Complex and Detailed Prompts"☆81Updated last year
- [CVPR 2023 (Highlight)] FAME-ViL: Multi-Tasking V+L Model for Heterogeneous Fashion Tasks☆55Updated last year