aimagelab / awesome-human-visual-attentionLinks
This repository contains a curated list of research papers and resources focusing on saliency and scanpath prediction, human attention, human visual search.
☆61Updated 6 months ago
Alternatives and similar repositories for awesome-human-visual-attention
Users that are interested in awesome-human-visual-attention are comparing it to the libraries listed below
Sorting:
- Official codebase for "Gazeformer: Scalable, Effective and Fast Prediction of Goal-Directed Human Attention" (CVPR 2023)☆41Updated last year
- ☆14Updated 9 months ago
- Learning Bottleneck Concepts in Image Classification (CVPR 2023)☆41Updated 2 years ago
- Official repository for "Video-FocalNets: Spatio-Temporal Focal Modulation for Video Action Recognition" [ICCV 2023]☆100Updated last year
- [CVPR 2023] Official repository of paper titled "Fine-tuned CLIP models are efficient video learners".☆298Updated last year
- Plotting heatmaps with the self-attention of the [CLS] tokens in the last layer.☆47Updated 3 years ago
- ☆23Updated last year
- [NeurIPS2022] Mind Reader: Reconstructing complex images from brain activities☆62Updated 2 years ago
- Official code repo for TCLR: Temporal Contrastive Learning for Video Representation [CVIU-2022]☆39Updated last year
- [ECCV 2024] - Improving Zero-shot Generalization of Learned Prompts via Unsupervised Knowledge Distillation☆64Updated 4 months ago
- ☆84Updated 2 years ago
- Sparse Linear Concept Embeddings☆120Updated 8 months ago
- [NeurIPS 2023] Align Your Prompts: Test-Time Prompting with Distribution Alignment for Zero-Shot Generalization☆108Updated last year
- ☆57Updated 3 years ago
- [CVPR 2023 & IJCV 2025] Positive-Augmented Contrastive Learning for Image and Video Captioning Evaluation☆64Updated 4 months ago
- repo for paper titled: Towards Realistic Zero-Shot Classification via Self Structural Semantic Alignment (AAAI'24 Oral)☆25Updated last year
- ☆62Updated 2 years ago
- Code implementation of our NeurIPS 2023 paper: Vocabulary-free Image Classification☆107Updated last year
- Official implementation of "Everything at Once - Multi-modal Fusion Transformer for Video Retrieval." CVPR 2022☆116Updated 3 years ago
- [BMVC 2023] Zero-shot Composed Text-Image Retrieval☆54Updated last year
- Code for "Modeling Multimodal Social Interactions: New Challenges and Baselines with Densely Aligned Representations" (CVPR 2024 Oral)☆17Updated last year
- Official repository for "Vita-CLIP: Video and text adaptive CLIP via Multimodal Prompting" [CVPR 2023]☆127Updated 2 years ago
- ☆54Updated last year
- [BMVC2022, IJCV2023, Best Student Paper, Spotlight] Official codes for the paper "In the Eye of Transformer: Global-Local Correlation for…☆29Updated 9 months ago
- Official Repository for VLLMs Provide Better Context for Emotion Understanding Through Common Sense Reasoning☆24Updated last year
- [ICLR 2024] FROSTER: Frozen CLIP is a Strong Teacher for Open-Vocabulary Action Recognition☆90Updated 10 months ago
- The PyTorch implementation for "DEAL: Disentangle and Localize Concept-level Explanations for VLMs" (ECCV 2024 Strong Double Blind)☆20Updated last year
- ☆69Updated last year
- [CVPRW-25 MMFM] Official repository of paper titled "How Good is my Video LMM? Complex Video Reasoning and Robustness Evaluation Suite fo…☆50Updated last year
- ☆60Updated last month