MCG-NJU / VideoMAELinks
[NeurIPS 2022 Spotlight] VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training
☆1,664Updated 2 years ago
Alternatives and similar repositories for VideoMAE
Users that are interested in VideoMAE are comparing it to the libraries listed below
Sorting:
- [CVPR 2023] VideoMAE V2: Scaling Video Masked Autoencoders with Dual Masking☆742Updated last year
- This is an official implementation for "Video Swin Transformers".☆1,624Updated 2 years ago
- ☆931Updated last year
- The official pytorch implementation of our paper "Is Space-Time Attention All You Need for Video Understanding?"☆1,822Updated last year
- [ICLR2022] official implementation of UniFormer☆895Updated last year
- Implementation of ViViT: A Video Vision Transformer☆556Updated 4 years ago
- VideoX: a collection of video cross-modal models☆1,056Updated last year
- [ECCV2024] Video Foundation Models & Data for Multimodal Understanding☆2,183Updated last month
- Code release for ActionFormer (ECCV 2022)☆536Updated last year
- [ECCV2024] VideoMamba: State Space Model for Efficient Video Understanding☆1,074Updated last year
- PyTorch implementation of a collections of scalable Video Transformer Benchmarks.☆305Updated 3 years ago
- This is the official implement of paper "ActionCLIP: A New Paradigm for Action Recognition"☆600Updated 2 years ago
- An official implementation for "CLIP4Clip: An Empirical Study of CLIP for End to End Video Clip Retrieval"☆1,022Updated last year
- Code Release for MViTv2 on Image Recognition.☆450Updated last year
- Implementation of CoCa, Contrastive Captioners are Image-Text Foundation Models, in Pytorch☆1,195Updated 2 years ago
- [ICCV2023] UniFormerV2: Spatiotemporal Learning by Arming Image ViTs with Video UniFormer☆339Updated last year
- Implementation of TimeSformer from Facebook AI, a pure attention-based solution for video classification☆727Updated 4 years ago
- Extract video features from raw videos using multiple GPUs. We support RAFT flow frames as well as S3D, I3D, R(2+1)D, VGGish, CLIP, and T…☆642Updated 11 months ago
- Video Swin Transformer - PyTorch☆265Updated 4 years ago
- Grounded Language-Image Pre-training☆2,569Updated 2 years ago
- Hiera: A fast, powerful, and simple hierarchical vision transformer.☆1,050Updated last year
- [ICLR 2023 Spotlight] Vision Transformer Adapter for Dense Predictions☆1,463Updated 7 months ago
- [CVPR 2024] Alpha-CLIP: A CLIP Model Focusing on Wherever You Want☆864Updated 6 months ago
- Temporal Action Detection & Weakly Supervised Temporal Action Detection & Temporal Action Proposal Generation☆566Updated last month
- EVA Series: Visual Representation Fantasies from BAAI☆2,639Updated last year
- Project Page for "LISA: Reasoning Segmentation via Large Language Model"☆2,566Updated 11 months ago
- [ICCV2023 Oral] Unmasked Teacher: Towards Training-Efficient Video Foundation Models☆347Updated last year
- Official Open Source code for "Masked Autoencoders As Spatiotemporal Learners"☆360Updated 2 weeks ago
- ❄️🔥 Visual Prompt Tuning [ECCV 2022] https://arxiv.org/abs/2203.12119☆1,209Updated 2 years ago
- (TPAMI 2024) A Survey on Open Vocabulary Learning☆985Updated last month