[AAAI2023] Symbolic Replay: Scene Graph as Prompt for Continual Learning on VQA Task (Oral)
☆41Mar 23, 2024Updated last year
Alternatives and similar repositories for CLVQA
Users that are interested in CLVQA are comparing it to the libraries listed below
Sorting:
- VQACL: A Novel Visual Question Answering Continual Learning Setting (CVPR'23)☆44Mar 28, 2024Updated last year
- [ICCV 2023] Label-Efficient Online Continual Object Detection in Streaming Video☆23Jan 8, 2024Updated 2 years ago
- Official Implementation for CVPR 2023 paper "Divide and Conquer: Answering Questions with Object Factorization and Compositional Reasonin…☆10Jun 16, 2024Updated last year
- Code for WACV 2023 paper "VLC-BERT: Visual Question Answering with Contextualized Commonsense Knowledge"☆21May 8, 2023Updated 2 years ago
- ☆73May 10, 2024Updated last year
- Affordance Grounding from Demonstration Video to Target Image (CVPR 2023)☆46Jul 26, 2024Updated last year
- Code for [CVPR 2025] ROICtrl: Boosting Instance Control for Visual Generation☆111Apr 16, 2025Updated 10 months ago
- Rich Visual Knowledge-based AugmentationNetwork for Visual Question Answering☆10Dec 6, 2019Updated 6 years ago
- [ICCV 2021] Official implementation of the paper "TRAR: Routing the Attention Spans in Transformers for Visual Question Answering"☆69Oct 11, 2021Updated 4 years ago
- This is the project page for the HOSNeRF☆16Dec 11, 2023Updated 2 years ago
- An Empirical Study of GPT-3 for Few-Shot Knowledge-Based VQA, AAAI 2022 (Oral)☆87Apr 10, 2022Updated 3 years ago
- HOSNeRF: Dynamic Human-Object-Scene Neural Radiance Fields from a Single Video☆68Dec 12, 2023Updated 2 years ago
- [ICLR2023] Towards Understanding and Mitigating Dimensional Collapse in Heterogeneous Federated Learning (https://arxiv.org/abs/2210.0022…☆40Jan 30, 2023Updated 3 years ago
- Code for MANO-GCN —— "Capturing Implicit Spatial Cues for Monocular 3D Hand Reconstruction" (ICME2021 Oral)☆13Jun 24, 2021Updated 4 years ago
- Code for ECCV 2022 paper “Learning with Recoverable Forgetting”☆21Jul 27, 2022Updated 3 years ago
- ☆18May 31, 2023Updated 2 years ago
- [CVPR 2024] How to Configure Good In-Context Sequence for Visual Question Answering☆21May 28, 2025Updated 9 months ago
- [ToMM2023] - AMC: Adaptive Multi-expert Collaborative Network for Text-guided Image Retrieval☆20Aug 30, 2024Updated last year
- [NeurIPS 2024] EvolveDirector: Approaching Advanced Text-to-Image Generation with Large Vision-Language Models.☆52Oct 14, 2024Updated last year
- Official Implementation of ReALFRED (ECCV'24)☆44Oct 11, 2024Updated last year
- ☆149Dec 7, 2024Updated last year
- Official implementation of the "Multimodal Parameter-Efficient Few-Shot Class Incremental Learning" paper☆24Apr 18, 2024Updated last year
- [Findings of EMNLP 2022] AssistSR: Task-oriented Video Segment Retrieval for Personal AI Assistant☆23Sep 11, 2023Updated 2 years ago
- ROSITA: Enhancing Vision-and-Language Semantic Alignments via Cross- and Intra-modal Knowledge Integration☆56Jun 13, 2023Updated 2 years ago
- visual question answering prompting recipes for large vision-language models☆28Sep 14, 2024Updated last year
- Learning Situation Hyper-Graphs for Video Question Answering☆22Feb 16, 2024Updated 2 years ago
- [IJCV 2025] Paragraph-to-Image Generation with Information-Enriched Diffusion Model☆106Mar 24, 2025Updated 11 months ago
- [CVPR 2025] DoraCycle: Domain-Oriented Adaptation of Unified Generative Model in Multimodal Cycles☆30May 13, 2025Updated 9 months ago
- ☆26Mar 20, 2023Updated 2 years ago
- CVPR 2022 Continual Learning in Computer Vision Workshop Challenge☆27Dec 15, 2022Updated 3 years ago
- MuKEA: Multimodal Knowledge Extraction and Accumulation for Knowledge-based Visual Question Answering☆100Mar 30, 2023Updated 2 years ago
- ☆140Mar 16, 2023Updated 2 years ago
- Code for NeurIPS 2022 paper “S-Prompts Learning with Pre-trained Transformers: An Occam’s Razor for Domain Incremental Learning“☆106Sep 26, 2024Updated last year
- CLEVR-X: A Visual Reasoning Dataset for Natural Language Explanations☆29Oct 27, 2023Updated 2 years ago
- [CVPR 2024] ViT-Lens: Towards Omni-modal Representations☆190Feb 3, 2025Updated last year
- ☆30Nov 29, 2023Updated 2 years ago
- PIC API☆25Sep 18, 2019Updated 6 years ago
- The Continual Learning in Multimodality Benchmark☆68Jun 24, 2023Updated 2 years ago
- OVMR: Open-Vocabulary Recognition with Multi-Modal References (CVPR24)☆35Jun 16, 2025Updated 8 months ago