Yuco-Z / Awesome-Multi-Modal-Dialog
[Paperlist] Awesome paper list of multimodal dialog, including methods, datasets and metrics
☆39Updated 2 months ago
Alternatives and similar repositories for Awesome-Multi-Modal-Dialog:
Users that are interested in Awesome-Multi-Modal-Dialog are comparing it to the libraries listed below
- [EMNLP'23 Oral] ReSee: Responding through Seeing Fine-grained Visual Knowledge in Open-domain Dialogue PyTorch Implementation☆12Updated last year
- Paper, dataset and code list for multimodal dialogue.☆20Updated 2 months ago
- This repo contains codes and instructions for baselines in the VLUE benchmark.☆41Updated 2 years ago
- [Paperlist] Awesome paper list of controllable text generation via latent auto-encoders. Contributions of any kind are welcome.☆51Updated 2 years ago
- Official code for our COLING 2022 paper: In-Context Learning for Empathetic Dialogue Generation☆19Updated 2 years ago
- This repository contains code to evaluate various multimodal large language models using different instructions across multiple multimoda…☆26Updated 3 weeks ago
- Code for our EMNLP-2022 paper: "Language Prior Is Not the Only Shortcut: A Benchmark for Shortcut Learning in VQA"☆38Updated 2 years ago
- VaLM: Visually-augmented Language Modeling. ICLR 2023.☆56Updated 2 years ago
- Implementation of our ACL2023 paper: Unifying Cross-Lingual and Cross-Modal Modeling Towards Weakly Supervised Multilingual Vision-Langua…☆19Updated last year
- ☆101Updated 2 years ago
- [ICML 2022] Code and data for our paper "IGLUE: A Benchmark for Transfer Learning across Modalities, Tasks, and Languages"☆49Updated 2 years ago
- [2024-ACL]: TextBind: Multi-turn Interleaved Multimodal Instruction-following in the Wildrounded Conversation☆47Updated last year
- Code for EMNLP 2022 paper “Distilled Dual-Encoder Model for Vision-Language Understanding”☆30Updated last year
- NAACL 2022: MCSE: Multimodal Contrastive Learning of Sentence Embeddings☆55Updated 9 months ago
- ☆49Updated 2 months ago
- Source code for the paper "Prefix Language Models are Unified Modal Learners"☆43Updated last year
- Official implementation of our EMNLP 2022 paper "CPL: Counterfactual Prompt Learning for Vision and Language Models"☆33Updated 2 years ago
- EMNLP2023 - InfoSeek: A New VQA Benchmark focus on Visual Info-Seeking Questions☆20Updated 10 months ago
- ☆17Updated last year
- [NeurIPS 2022 Workshop] A Case Study with Negated Prompts using T0 (3B, 11B), InstructGPT (350M-175B), GPT-3 (350M - 175B) & OPT (125M - …☆24Updated 2 years ago
- Dialogue Planning via Brownian Bridge Stochastic Process for Goal-directed Proactive Dialogue (ACL Findings 2023)☆22Updated last year
- MultiInstruct: Improving Multi-Modal Zero-Shot Learning via Instruction Tuning☆136Updated last year
- ☆15Updated 2 years ago
- ☆65Updated 6 years ago
- ☆152Updated 3 years ago
- [CVPR 2021] Counterfactual VQA: A Cause-Effect Look at Language Bias☆120Updated 3 years ago
- [NeurIPS 2022] Non-Linguistic Supervision for Contrastive Learning of Sentence Embeddings☆21Updated 2 years ago
- Source code and data for Things not Written in Text: Exploring Spatial Commonsense from Visual Signals (ACL2022 main conference paper).☆19Updated 2 years ago
- This is the GPT2 baseline for ProtoQA☆12Updated 3 years ago
- Recent Advances in Visual Dialog☆30Updated 2 years ago