YehLi / xmodaler
X-modaler is a versatile and high-performance codebase for cross-modal analytics(e.g., image captioning, video captioning, vision-language pre-training, visual question answering, visual commonsense reasoning, and cross-modal retrieval).
☆1,030Updated last year
Related projects ⓘ
Alternatives and complementary repositories for xmodaler
- [CVPR 2021 Best Student Paper Honorable Mention, Oral] Official PyTorch code for ClipBERT, an efficient framework for end-to-end learning…☆704Updated last year
- VideoX: a collection of video cross-modal models☆980Updated 5 months ago
- Implementation of 'X-Linear Attention Networks for Image Captioning' [CVPR 2020]☆271Updated 3 years ago
- An official implementation for "CLIP4Clip: An Empirical Study of CLIP for End to End Video Clip Retrieval"☆879Updated 7 months ago
- A curated list of deep learning resources for video-text retrieval.☆592Updated last year
- An official implementation for " UniVL: A Unified Video and Language Pre-Training Model for Multimodal Understanding and Generation"☆339Updated 3 months ago
- X-VLM: Multi-Grained Vision Language Pre-Training (ICML 2022)☆448Updated last year
- 【CVPR'2023 Highlight & TPAMI】Cap4Video: What Can Auxiliary Captions Do for Text-Video Retrieval?☆238Updated 2 months ago
- This repository focus on Image Captioning & Video Captioning & Seq-to-Seq Learning & NLP☆413Updated last year
- Multi-Modal Transformer for Video Retrieval☆258Updated last month
- ☆231Updated last year
- A PyTorch reimplementation of bottom-up-attention models☆292Updated 2 years ago
- The Paper List of Large Multi-Modality Model, Parameter-Efficient Finetuning, Vision-Language Pretraining, Conventional Image-Text Matchi…☆399Updated 4 months ago
- Video embeddings for retrieval with natural language queries☆335Updated last year
- Video Grounding and Captioning☆323Updated 3 years ago
- Code accompanying the paper "Fine-grained Video-Text Retrieval with Hierarchical Graph Reasoning".☆209Updated 4 years ago
- METER: A Multimodal End-to-end TransformER Framework☆362Updated last year
- Recent Advances in Vision and Language Pre-training (VLP)☆288Updated last year
- [ICLR 2022] code for "How Much Can CLIP Benefit Vision-and-Language Tasks?" https://arxiv.org/abs/2107.06383☆401Updated 2 years ago
- [NeurIPS 2022 Spotlight] VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training☆1,371Updated 11 months ago
- Research code for EMNLP 2020 paper "HERO: Hierarchical Encoder for Video+Language Omni-representation Pre-training"☆230Updated 3 years ago
- Meshed-Memory Transformer for Image Captioning. CVPR 2020☆519Updated last year
- Research code for CVPR 2022 paper "SwinBERT: End-to-End Transformers with Sparse Attention for Video Captioning"☆237Updated 2 years ago
- Multi-modality pre-training☆471Updated 6 months ago
- Oscar and VinVL☆1,038Updated last year
- project page for VinVL☆350Updated last year
- [ECCV2024] Grounded Multimodal Large Language Model with Localized Visual Tokenization☆558Updated 5 months ago
- End-to-End Dense Video Captioning with Parallel Decoding (ICCV 2021)☆208Updated 10 months ago
- awesome grounding: A curated list of research papers in visual grounding☆1,026Updated last year
- [CVPR 2023] Prompt, Generate, then Cache: Cascade of Foundation Models makes Strong Few-shot Learners☆349Updated last year
- Code for ALBEF: a new vision-language pre-training method☆1,557Updated 2 years ago