Chiaraplizz / ARGO1M-What-can-a-cookLinks
☆10Updated last year
Alternatives and similar repositories for ARGO1M-What-can-a-cook
Users that are interested in ARGO1M-What-can-a-cook are comparing it to the libraries listed below
Sorting:
- The implementation of CVPR2021 paper Temporal Query Networks for Fine-grained Video Understanding☆62Updated 3 years ago
- This repository contains the Adverbs in Recipes (AIR) dataset and the code published at the CVPR 23 paper: "Learning Action Changes by Me…☆13Updated 2 years ago
- CVPR2022☆21Updated 2 years ago
- ☆26Updated last year
- Code implementation for our ECCV, 2022 paper titled "My View is the Best View: Procedure Learning from Egocentric Videos"☆28Updated last year
- HT-Step is a large-scale article grounding dataset of temporal step annotations on how-to videos☆18Updated last year
- Implementation of paper 'Helping Hands: An Object-Aware Ego-Centric Video Recognition Model'☆33Updated last year
- Temporal Action Localization Visualization Tool (TALVT) is a Javascript based simple visualization tool to visualize the outcomes of the …☆28Updated 4 years ago
- Codebase for "Revisiting spatio-temporal layouts for compositional action recognition" (Oral at BMVC 2021).☆26Updated 3 years ago
- Official code for "Disentangling Visual Embeddings for Attributes and Objects" Published at CVPR 2022☆35Updated last year
- Official code implemtation of paper AntGPT: Can Large Language Models Help Long-term Action Anticipation from Videos?☆21Updated 8 months ago
- Code for Static and Dynamic Concepts for Self-supervised Video Representation Learning.☆10Updated 2 years ago
- Implementation of "With a Little Help from my Temporal Context: Multimodal Egocentric Action Recognition, BMVC, 2021" in PyTorch☆19Updated 3 years ago
- CVPR2022:Learning from Untrimmed Videos: Self-Supervised Video Representation Learning with Hierarchical Consistency☆17Updated 2 years ago
- Code accompanying Ego-Exo: Transferring Visual Representations from Third-person to First-person Videos (CVPR 2021)☆33Updated 3 years ago
- MAtch, eXpand and Improve: Unsupervised Finetuning for Zero-Shot Action Recognition with Language Knowledge (ICCV 2023)☆30Updated last year
- Official implementation of "Test-Time Zero-Shot Temporal Action Localization", CVPR 2024☆60Updated 8 months ago
- The champion solution for Ego4D Natural Language Queries Challenge in CVPR 2023☆17Updated last year
- [ Arxiv 2023 ] This repository contains the code for "MUPPET: Multi-Modal Few-Shot Temporal Action Detection"☆15Updated last year
- This is the official implementation of Elaborative Rehearsal for Zero-shot Action Recognition (ICCV2021)☆36Updated 3 years ago
- Future Transformer for Long-term Action Anticipation (CVPR 2022)☆49Updated 2 years ago
- BEAR: a new BEnchmark on video Action Recognition☆44Updated last year
- X-MIC: Cross-Modal Instance Conditioning for Egocentric Action Generalization, CVPR 2024☆11Updated 6 months ago
- [CVPR'22 Oral] Temporal Alignment Networks for Long-term Video. Tengda Han, Weidi Xie, Andrew Zisserman.☆118Updated last year
- [ICLR 2023] Temporal Alignment Representations with Contrastive Learning☆26Updated 2 years ago
- [ICCV 2023] How Much Temporal Long-Term Context is Needed for Action Segmentation?☆45Updated 11 months ago
- Code release for "EgoVLPv2: Egocentric Video-Language Pre-training with Fusion in the Backbone" [ICCV, 2023]☆98Updated 11 months ago
- Official implementation of "HowToCaption: Prompting LLMs to Transform Video Annotations at Scale." ECCV 2024☆53Updated 8 months ago
- ICCV2023: Disentangling Spatial and Temporal Learning for Efficient Image-to-Video Transfer Learning☆41Updated last year
- Code and data release for the paper "Learning Fine-grained View-Invariant Representations from Unpaired Ego-Exo Videos via Temporal Align…☆17Updated last year