[ICCV2023] UniFormerV2: Spatiotemporal Learning by Arming Image ViTs with Video UniFormer
☆339Apr 2, 2024Updated last year
Alternatives and similar repositories for UniFormerV2
Users that are interested in UniFormerV2 are comparing it to the libraries listed below
Sorting:
- [ICLR2022] official implementation of UniFormer☆896Mar 29, 2024Updated last year
- [ICCV2023 Oral] Unmasked Teacher: Towards Training-Efficient Video Foundation Models☆347May 27, 2024Updated last year
- [ECCV2024] Video Foundation Models & Data for Multimodal Understanding☆2,201Dec 15, 2025Updated 2 months ago
- [CVPR 2023] VideoMAE V2: Scaling Video Masked Autoencoders with Dual Masking☆752Oct 8, 2024Updated last year
- [NeurIPS 2022 Spotlight] VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training☆1,681Dec 8, 2023Updated 2 years ago
- An unofficial implementation of TubeViT in "Rethinking Video ViTs: Sparse Video Tubes for Joint Image and Video Learning"☆94Sep 13, 2024Updated last year
- ☆181Aug 20, 2022Updated 3 years ago
- [CVPR2024 Highlight][VideoChatGPT] ChatGPT with video understanding! And many more supported LMs such as miniGPT4, StableLM, and MOSS.☆3,334Jan 18, 2025Updated last year
- OpenMMLab's Next Generation Video Understanding Toolbox and Benchmark☆4,930Aug 14, 2024Updated last year
- Champion Solutions repository for Perception Test challenges in ICCV2023 workshop.☆14Oct 18, 2023Updated 2 years ago
- This is an official implementation for "Video Swin Transformers".☆1,632Mar 8, 2023Updated 2 years ago
- EVA Series: Visual Representation Fantasies from BAAI☆2,647Aug 1, 2024Updated last year
- Code release for ActionFormer (ECCV 2022)☆541Apr 11, 2024Updated last year
- [NeurIPS2022] This is the official implementation of the paper "Expediting Large-Scale Vision Transformer for Dense Prediction without Fi…☆86Oct 29, 2023Updated 2 years ago
- [ICCV 2023] MGMAE: Motion Guided Masking for Video Masked Autoencoding☆26Oct 16, 2023Updated 2 years ago
- [ICLR2026] VideoChat-Flash: Hierarchical Compression for Long-Context Video Modeling☆510Nov 18, 2025Updated 3 months ago
- PySlowFast: video understanding codebase from FAIR for reproducing state-of-the-art video models.☆7,297Feb 19, 2026Updated last week
- ☆18Oct 6, 2022Updated 3 years ago
- ☆285Aug 14, 2025Updated 6 months ago
- [ECCV2024] VideoMamba: State Space Model for Efficient Video Understanding☆1,081Jul 6, 2024Updated last year
- VideoX: a collection of video cross-modal models☆1,061Jun 3, 2024Updated last year
- The official pytorch implementation of our paper "Is Space-Time Attention All You Need for Video Understanding?"☆1,830Apr 9, 2024Updated last year
- An official implementation for "CLIP4Clip: An Empirical Study of CLIP for End to End Video Clip Retrieval"☆1,025Apr 12, 2024Updated last year
- [CVPR 2023] Official repository of paper titled "Fine-tuned CLIP models are efficient video learners".☆304Apr 3, 2024Updated last year
- [ICLR 2023] Official implementation of the paper "DINO: DETR with Improved DeNoising Anchor Boxes for End-to-End Object Detection"☆2,752Jul 31, 2024Updated last year
- [CVPR 2023 Highlight] InternImage: Exploring Large-Scale Vision Foundation Models with Deformable Convolutions☆2,793Mar 25, 2025Updated 11 months ago
- ☆110Dec 23, 2022Updated 3 years ago
- Position sensitive PreciseRoIPooling without roi coordinates gradient backward☆16Aug 2, 2018Updated 7 years ago
- Task Preference Optimization: Improving Multimodal Large Language Models with Vision Task Alignment☆64Jul 22, 2025Updated 7 months ago
- [ICCV 2023] You Only Look at One Partial Sequence☆343Oct 21, 2023Updated 2 years ago
- General Vision Benchmark, GV-B, a project from OpenGVLab☆188Feb 23, 2022Updated 4 years ago
- ☆49Nov 12, 2022Updated 3 years ago
- [CVPR2023] All in One: Exploring Unified Video-Language Pre-training☆281Mar 25, 2023Updated 2 years ago
- A deep learning library for video understanding research.☆3,544Jan 12, 2026Updated last month
- Scenic: A Jax Library for Computer Vision Research and Beyond☆3,766Feb 18, 2026Updated last week
- Code release for "Learning Video Representations from Large Language Models"☆536Oct 1, 2023Updated 2 years ago
- CLIP Itself is a Strong Fine-tuner: Achieving 85.7% and 88.0% Top-1 Accuracy with ViT-B and ViT-L on ImageNet☆224Dec 16, 2022Updated 3 years ago
- Official repository for "Self-Supervised Video Transformer" (CVPR'22)☆108Jun 26, 2024Updated last year
- [CVPR'23] AdaMAE: Adaptive Masking for Efficient Spatiotemporal Learning with Masked Autoencoders☆84Feb 2, 2024Updated 2 years ago