nguyennm1024 / OSCaRLinks
π₯π₯π₯ Object State Description & Change Detection
β10Updated last year
Alternatives and similar repositories for OSCaR
Users that are interested in OSCaR are comparing it to the libraries listed below
Sorting:
- Code and data release for the paper "Learning Object State Changes in Videos: An Open-World Perspective" (CVPR 2024)β33Updated 9 months ago
- Code and data release for the paper "Learning Fine-grained View-Invariant Representations from Unpaired Ego-Exo Videos via Temporal Alignβ¦β17Updated last year
- Visual Representation Learning with Stochastic Frame Prediction (ICML 2024)β21Updated 7 months ago
- Official code implemtation of paper AntGPT: Can Large Language Models Help Long-term Action Anticipation from Videos?β21Updated 9 months ago
- IMProv: Inpainting-based Multimodal Prompting for Computer Vision Tasksβ57Updated 9 months ago
- Code for Look for the Change paper published at CVPR 2022β36Updated 2 years ago
- ChangeIt dataset with more than 2600 hours of video with state-changing actions published at CVPR 2022β11Updated 3 years ago
- Code implementation for our ECCV, 2022 paper titled "My View is the Best View: Procedure Learning from Egocentric Videos"β28Updated last year
- Code implementation for paper titled "HOI-Ref: Hand-Object Interaction Referral in Egocentric Vision"β27Updated last year
- Code for the paper "Multi-Task Learning of Object States and State-Modifying Actions from Web Videos" published in TPAMIβ11Updated last year
- Affordance Grounding from Demonstration Video to Target Image (CVPR 2023)β44Updated 11 months ago
- HT-Step is a large-scale article grounding dataset of temporal step annotations on how-to videosβ19Updated last year
- [ICLR 2024 Poster] SCHEMA: State CHangEs MAtter for Procedure Planning in Instructional Videosβ18Updated 7 months ago
- Official implementation of the CVPR'24 paper [Adaptive Slot Attention: Object Discovery with Dynamic Slot Number]β50Updated 5 months ago
- [ECCV2024, Oral, Best Paper Finalist] This is the official implementation of the paper "LEGO: Learning EGOcentric Action Frame Generationβ¦β37Updated 4 months ago
- [AAAI2023] Symbolic Replay: Scene Graph as Prompt for Continual Learning on VQA Task (Oral)β39Updated last year
- Ego4D Goal-Step: Toward Hierarchical Understanding of Procedural Activities (NeurIPS 2023)β44Updated last year
- This repository is the official implementation of Improving Object-centric Learning With Query Optimizationβ50Updated 2 years ago
- Official code for Slot-Transformer for Videos (STEVE)β49Updated 2 years ago
- Code for the VOST datasetβ26Updated last year
- [NeurIPS 2023] Self-supervised Object-Centric Learning for Videosβ27Updated 6 months ago
- [CVPR 2024] Data and benchmark code for the EgoExoLearn datasetβ61Updated 9 months ago
- A repo for processing the raw hand object detections to produce releasable pickles + library for using theseβ37Updated 8 months ago
- β41Updated last year
- This repository contains the Adverbs in Recipes (AIR) dataset and the code published at the CVPR 23 paper: "Learning Action Changes by Meβ¦β13Updated 2 years ago
- Code for NeurIPS 2022 Datasets and Benchmarks paper - EgoTaskQA: Understanding Human Tasks in Egocentric Videos.β33Updated 2 years ago
- Self-supervised algorithm for learning representations from ego-centric video data. Code is tested on EPIC-Kitchens-100 and Ego4D in PyToβ¦β12Updated 2 years ago
- Official Code for Neural Systematic Binderβ33Updated 2 years ago
- β19Updated last year
- β71Updated 6 months ago