Imageomics / bioclip
This is the repository for the BioCLIP model and the TreeOfLife-10M dataset [CVPR'24 Oral, Best Student Paper].
☆200Updated last month
Alternatives and similar repositories for bioclip:
Users that are interested in bioclip are comparing it to the libraries listed below
- This is an official implementation for [ICLR'24] INTR: Interpretable Transformer for Fine-grained Image Classification.☆49Updated last year
- This repo contains the evaluation code for the INQUIRE benchmark☆46Updated 4 months ago
- Official Implementation of the CrossMAE paper: Rethinking Patch Dependence for Masked Autoencoders☆107Updated 2 weeks ago
- Object Recognition as Next Token Prediction (CVPR 2024 Highlight)☆175Updated 4 months ago
- [NeurIPS 2023] This repo contains the code for our paper Convolutions Die Hard: Open-Vocabulary Segmentation with Single Frozen Convoluti…☆313Updated last year
- [NeurIPS 2024] Official implementation of the paper "Interfacing Foundation Models' Embeddings"☆123Updated 8 months ago
- [CVPR 2024] Code release for "Unsupervised Universal Image Segmentation"☆198Updated 11 months ago
- [CVPR24] Official Implementation of GEM (Grounding Everything Module)☆117Updated 2 weeks ago
- ☆201Updated last year
- [CVPR 24] The repository provides code for running inference and training for "Segment and Caption Anything" (SCA) , links for downloadin…☆220Updated 6 months ago
- Official repository of paper "Subobject-level Image Tokenization"☆69Updated 3 weeks ago
- [NeurIPS2023] Code release for "Hierarchical Open-vocabulary Universal Image Segmentation"☆286Updated last year
- Connecting segment-anything's output masks with the CLIP model; Awesome-Segment-Anything-Works☆191Updated 6 months ago
- Official implementation of 'CLIP-DINOiser: Teaching CLIP a few DINO tricks' paper.☆244Updated 6 months ago
- [NeurIPS 2024] Code release for "Segment Anything without Supervision"☆461Updated 6 months ago
- This is Pytorch Implementation Code for adding new features in code of Segment-Anything. Here, the features support batch-input on the fu…☆154Updated last year
- Holds code for our CVPR'23 tutorial: All Things ViTs: Understanding and Interpreting Attention in Vision.☆186Updated last year
- [NeurIPS 2024] MoVA: Adapting Mixture of Vision Experts to Multimodal Context☆154Updated 7 months ago
- [ECCV'24] Official Implementation of SemiVL: Semi-Supervised Semantic Segmentation with Vision-Language Guidance☆128Updated last week
- [NeurIPS 2023] This repository includes the official implementation of our paper "An Inverse Scaling Law for CLIP Training"☆314Updated 10 months ago
- Code for experiments for "ConvNet vs Transformer, Supervised vs CLIP: Beyond ImageNet Accuracy"☆101Updated 7 months ago
- [CVPR2024] ViP-LLaVA: Making Large Multimodal Models Understand Arbitrary Visual Prompts☆319Updated 9 months ago
- Official implementation of "Describing Differences in Image Sets with Natural Language" (CVPR 2024 Oral)☆118Updated last year
- Reproducible scaling laws for contrastive language-image learning (https://arxiv.org/abs/2212.07143)☆164Updated last year
- NeurIPS 2024 Track on Datasets and Benchmarks (Spotlight)☆29Updated last week
- WildlifeDatasets: An open-source toolkit for animal re-identification☆100Updated last week
- VisionLLaMA: A Unified LLaMA Backbone for Vision Tasks☆385Updated 9 months ago
- Projects based on SigLIP (Zhai et. al, 2023) and Hugging Face transformers integration 🤗☆231Updated 2 months ago
- Official implementation of SCLIP: Rethinking Self-Attention for Dense Vision-Language Inference☆154Updated 6 months ago
- When do we not need larger vision models?☆388Updated 2 months ago