☆236Aug 5, 2025Updated 7 months ago
Alternatives and similar repositories for CLIP-rsicd
Users that are interested in CLIP-rsicd are comparing it to the libraries listed below
Sorting:
- Datasets for remote sensing images (Paper:Exploring Models and Data for Remote Sensing Image Caption Generation)☆229Nov 28, 2021Updated 4 years ago
- ☆18Nov 23, 2022Updated 3 years ago
- Collection of Remote Sensing Vision-Language Models☆142May 13, 2024Updated last year
- Source code of paper "Remote Sensing Cross-Modal Image-Text Retrieval Based on Global and Local Information"☆72Oct 25, 2023Updated 2 years ago
- 🛰️ Official repository of paper "RemoteCLIP: A Vision Language Foundation Model for Remote Sensing" (IEEE TGRS)☆525Jun 27, 2024Updated last year
- RS5M: a large-scale vision language dataset for remote sensing [TGRS]☆301Mar 17, 2025Updated last year
- Parameter-Efficient Transfer Learning for Remote Sensing Image-Text Retrieval, 2023☆29Jan 14, 2024Updated 2 years ago
- A list of awesome remote sensing image captioning resources☆122Mar 14, 2026Updated last week
- RSVG: Exploring Data and Model for Visual Grounding on Remote Sensing Data, 2022☆170Dec 10, 2025Updated 3 months ago
- [IEEE TGRS 2022 🔥] Remote Sensing Image Change Captioning With Dual-Branch Transformers: A New Method and a Large Scale Dataset☆138Sep 16, 2025Updated 6 months ago
- modified datasets for remote sensing image caption☆11Apr 23, 2019Updated 6 years ago
- Official repo for "SkyScript: A Large and Semantically Diverse Vision-Language Dataset for Remote Sensing"☆198Dec 10, 2024Updated last year
- A PyTorch implementation of "GeoSynth: Contextually-Aware High-Resolution Satellite Image Synthesis"☆111Nov 29, 2024Updated last year
- Multimodal Large Language Models for Remote Sensing (RS-MLLMs): A Survey☆373Jan 12, 2026Updated 2 months ago
- This is tensorflow 2.2 based SCAMET framework for remote sensing image captioning.☆13Aug 10, 2023Updated 2 years ago
- Awesome-Remote-Sensing-Vision-Language-Models☆192Apr 27, 2024Updated last year
- Official Code for “PriorCLIP: Visual Prior Guided Vision-Language Model for Remote Sensing Image-Text Retrieval”☆26Dec 19, 2025Updated 3 months ago
- ☆13Sep 28, 2024Updated last year
- Remote sensing Image Captioning is a special case of Image Captioning which solves the difficulties in processing the remote sensing imag…☆11Jun 16, 2021Updated 4 years ago
- remote sensing image classification and image caption by PyTorch☆60Jul 24, 2019Updated 6 years ago
- CLIP (Contrastive Language–Image Pre-training) for Italian☆185May 11, 2023Updated 2 years ago
- [ICLRW 2024] Efficient Remote Sensing with Harmonized Transfer Learning and Modality Alignment☆63Jul 18, 2024Updated last year
- [CVPR 2024 🔥] GeoChat, the first grounded Large Vision Language Model for Remote Sensing☆698Nov 28, 2024Updated last year
- Robust fine-tuning of zero-shot models☆760Apr 29, 2022Updated 3 years ago
- An open source implementation of CLIP.☆13,528Mar 12, 2026Updated last week
- This repository contains code to reproduce the experiments in the preprint "MMEarth: Exploring Multi-Modal Pretext Tasks For Geospatial R…☆63Jan 6, 2026Updated 2 months ago
- [IEEE TIP 2023] Txt2Img-MHN: Remote Sensing Image Generation from Text Using Modern Hopfield Networks☆82Oct 19, 2023Updated 2 years ago
- This is the official code for "Bootstrapping Interactive Image-Text Alignment for Remote Sensing Image Captioning"☆35Dec 24, 2024Updated last year
- ☆145Dec 11, 2025Updated 3 months ago
- A PyTorch Lightning solution to training OpenAI's CLIP from scratch.☆720Apr 15, 2022Updated 3 years ago
- The source code of AMFMN and the dataset RSITMD☆217Oct 25, 2023Updated 2 years ago
- Multiform Ensemble Self-Supervised Learning for Few-Shot Remote Sensing Scene Classification☆13Mar 10, 2023Updated 3 years ago
- S-CLIP: Semi-supervised Vision-Language Pre-training using Few Specialist Captions☆51May 26, 2023Updated 2 years ago
- Official PyTorch implementation for Hypersphere-Based Remote Sensing Cross-Modal Text–Image Retrieval via Curriculum Learning.☆16Aug 10, 2024Updated last year
- ☆24Aug 16, 2024Updated last year
- CLIP Itself is a Strong Fine-tuner: Achieving 85.7% and 88.0% Top-1 Accuracy with ViT-B and ViT-L on ImageNet☆223Dec 16, 2022Updated 3 years ago
- Architectures for Remote Sensing Image Captioning Thesis☆12Jul 30, 2024Updated last year
- CLIP-Art: Contrastive Pre-training for Fine-Grained Art Classification - 4th Workshop on Computer Vision for Fashion, Art, and Design☆28May 2, 2022Updated 3 years ago
- ☆97Dec 15, 2025Updated 3 months ago