CMU-MultiComp-Lab / mmml-courseLinks
☆91Updated last year
Alternatives and similar repositories for mmml-course
Users that are interested in mmml-course are comparing it to the libraries listed below
Sorting:
- ☆36Updated last year
- ☆29Updated last year
- [TMLR 2022] High-Modality Multimodal Transformer☆115Updated 7 months ago
- This repository holds code and other relevant files for the NeurIPS 2022 tutorial: Foundational Robustness of Foundation Models.☆70Updated 2 years ago
- [ICLR 2023] MultiViz: Towards Visualizing and Understanding Multimodal Models☆96Updated 10 months ago
- https://slds-lmu.github.io/seminar_multimodal_dl/☆169Updated 2 years ago
- A Survey on multimodal learning research.☆328Updated last year
- Visual Language Transformer Interpreter - An interactive visualization tool for interpreting vision-language transformers☆93Updated last year
- A curated list of vision-and-language pre-training (VLP). :-)☆59Updated 2 years ago
- Video descriptions of research papers relating to foundation models and scaling☆31Updated 2 years ago
- Reading list for Multimodal Large Language Models☆67Updated last year
- Open source code for AAAI 2023 Paper "BridgeTower: Building Bridges Between Encoders in Vision-Language Representation Learning"☆165Updated last year
- This repository provides a comprehensive collection of research papers focused on multimodal representation learning, all of which have b…☆76Updated last week
- ICLR 2023 Paper submission analysis from https://openreview.net/group?id=ICLR.cc/2023/Conference☆105Updated 2 years ago
- A curated list of the papers, repositories, tutorials, and anythings related to the large language models for tools☆67Updated last year
- The repository collects many various multi-modal transformer architectures, including image transformer, video transformer, image-languag…☆230Updated 2 years ago
- ☆97Updated 2 years ago
- [T-PAMI] A curated list of self-supervised multimodal learning resources.☆261Updated 10 months ago
- Collection of Tools and Papers related to Adapters / Parameter-Efficient Transfer Learning/ Fine-Tuning☆192Updated last year
- In-the-wild Question Answering☆15Updated 2 years ago
- Holistic evaluation of multimodal foundation models☆47Updated 10 months ago
- Research Trends in LLM-guided Multimodal Learning.☆358Updated last year
- Coarse-to-Fine Vision-Language Pre-training with Fusion in the Backbone☆129Updated last year
- AI Alignment: A Comprehensive Survey☆135Updated last year
- Residual Prompt Tuning: a method for faster and better prompt tuning.☆54Updated 2 years ago
- Mind the Gap: Understanding the Modality Gap in Multi-modal Contrastive Representation Learning☆158Updated 2 years ago
- ☆81Updated 10 months ago
- MLLM-Bench: Evaluating Multimodal LLMs with Per-sample Criteria☆69Updated 8 months ago
- [NeurIPS 2023] Text data, code and pre-trained models for paper "Improving CLIP Training with Language Rewrites"☆282Updated last year
- PyTorch code for "VL-Adapter: Parameter-Efficient Transfer Learning for Vision-and-Language Tasks" (CVPR2022)☆205Updated 2 years ago