PalAvik / hycoclipLinks
Code for the paper "Compositional Entailment Learning for Hyperbolic Vision-Language Models".
☆95Updated 6 months ago
Alternatives and similar repositories for hycoclip
Users that are interested in hycoclip are comparing it to the libraries listed below
Sorting:
- [ICLR 2024] Test-Time RL with CLIP Feedback for Vision-Language Models.☆97Updated 2 months ago
- Diffusion-TTA improves pre-trained discriminative models such as image classifiers or segmentors using pre-trained generative models.☆79Updated last year
- [CVPR 2025] FLAIR: VLM with Fine-grained Language-informed Image Representations☆128Updated 4 months ago
- [ECCV 2024] Mind the Interference: Retaining Pre-trained Knowledge in Parameter Efficient Continual Learning of Vision-Language Models☆56Updated last year
- The official pytorch implemention of our CVPR-2024 paper "MMA: Multi-Modal Adapter for Vision-Language Models".☆94Updated 8 months ago
- cliptrase☆47Updated last year
- ☆22Updated 7 months ago
- FreeDA: Training-Free Open-Vocabulary Segmentation with Offline Diffusion-Augmented Prototype Generation (CVPR 2024)☆48Updated last year
- [CVPR 2025] Understanding Fine-tuning CLIP for Open-vocabulary Semantic Segmentation in Hyperbolic Space☆31Updated 5 months ago
- AlignCLIP: Improving Cross-Modal Alignment in CLIP (ICLR 2025)☆52Updated 10 months ago
- Pytorch Implementation for CVPR 2024 paper: Learn to Rectify the Bias of CLIP for Unsupervised Semantic Segmentation☆56Updated 4 months ago
- Official Implementation of the ECCV 2024 Paper: "CLAP: Isolating Content from Style through Contrastive Learning with Augmented Prompts"☆54Updated 2 months ago
- [CVPR 2024] Improving language-visual pretraining efficiency by perform cluster-based masking on images.☆29Updated last year
- [ICLR2025] This repository is the official implementation of our Autoregressive Pretraining with Mamba in Vision☆89Updated 7 months ago
- [ECCV 2022] What to Hide from Your Students: Attention-Guided Masked Image Modeling☆74Updated last year
- [CVPR 2025] COSMOS: Cross-Modality Self-Distillation for Vision Language Pre-training☆37Updated 9 months ago
- [ICLR 2025] - Cross the Gap: Exposing the Intra-modal Misalignment in CLIP via Modality Inversion☆58Updated last month
- official implementation of "Interpreting CLIP's Image Representation via Text-Based Decomposition"☆234Updated 7 months ago
- ☆22Updated last year
- Official Implementation of "Read-only Prompt Optimization for Vision-Language Few-shot Learning", ICCV 2023☆54Updated 2 years ago
- [ICCV25 Oral] Token Activation Map to Visually Explain Multimodal LLMs☆145Updated 2 weeks ago
- [ECCV2024] ProxyCLIP: Proxy Attention Improves CLIP for Open-Vocabulary Segmentation☆112Updated 9 months ago
- Official implementation of CVPR 2024 paper "Prompt Learning via Meta-Regularization".☆31Updated 9 months ago
- Easy wrapper for inserting LoRA layers in CLIP.☆40Updated last year
- Detail-Oriented CLIP for Fine-Grained Tasks (ICLR SSI-FM 2025)☆56Updated 9 months ago
- Official implementation of ResCLIP: Residual Attention for Training-free Dense Vision-language Inference☆55Updated 2 months ago
- [ICML'25] Kernel-based Unsupervised Embedding Alignment for Enhanced Visual Representation in Vision-language Models☆19Updated 3 months ago
- [ICLR 2024] SemiReward: A General Reward Model for Semi-supervised Learning☆76Updated last month
- Official implementation of SCLIP: Rethinking Self-Attention for Dense Vision-Language Inference☆180Updated last year
- [ECCV 2024] Official repository for "DataDream: Few-shot Guided Dataset Generation"☆46Updated last year