KastanDay / video-pretrained-transformerLinks
Multi-model video-to-text by combining embeddings from Flan-T5 + CLIP + Whisper + SceneGraph. The 'backbone LLM' is pre-trained from scratch on YouTube (YT-1B dataset).
☆52Updated 2 years ago
Alternatives and similar repositories for video-pretrained-transformer
Users that are interested in video-pretrained-transformer are comparing it to the libraries listed below
Sorting:
- VideoLLM: Modeling Video Sequence with Large Language Models☆158Updated 2 years ago
- [ICLR2024] Codes and Models for COSA: Concatenated Sample Pretrained Vision-Language Foundation Model☆43Updated 10 months ago
- Democratization of "PaLI: A Jointly-Scaled Multilingual Language-Image Model"☆91Updated last year
- Code for our ACL 2025 paper "Language Repository for Long Video Understanding"☆32Updated last year
- [TMLR23] Official implementation of UnIVAL: Unified Model for Image, Video, Audio and Language Tasks.☆231Updated last year
- [CVPR 2023] HierVL Learning Hierarchical Video-Language Embeddings☆46Updated 2 years ago
- Graph learning framework for long-term video understanding☆68Updated 3 months ago
- Fine-tuning "ImageBind One Embedding Space to Bind Them All" with LoRA☆192Updated last year
- [NeurIPS 2023 D&B] VidChapters-7M: Video Chapters at Scale☆198Updated last year
- Implementation of the model: "(MC-ViT)" from the paper: "Memory Consolidation Enables Long-Context Video Understanding"☆23Updated last week
- Code release for "EgoVLPv2: Egocentric Video-Language Pre-training with Fusion in the Backbone" [ICCV, 2023]☆100Updated last year
- Implementation of PALI3 from the paper PALI-3 VISION LANGUAGE MODELS: SMALLER, FASTER, STRONGER"☆145Updated last week
- PG-Video-LLaVA: Pixel Grounding in Large Multimodal Video Models☆259Updated 3 months ago
- ☆20Updated 5 months ago
- [CVPR 2024] ViT-Lens: Towards Omni-modal Representations☆183Updated 9 months ago
- LAVIS - A One-stop Library for Language-Vision Intelligence