eujhwang / meme-capLinks
☆39Updated 2 years ago
Alternatives and similar repositories for meme-cap
Users that are interested in meme-cap are comparing it to the libraries listed below
Sorting:
- Research code for pixel-based encoders of language (PIXEL)☆337Updated this week
- ☆114Updated last year
- ☆22Updated 3 months ago
- PyTorch code for "Unifying Vision-and-Language Tasks via Text Generation" (ICML 2021)☆371Updated last year
- Corpus to accompany: "Do Androids Laugh at Electric Sheep? Humor "Understanding" Benchmarks from The New Yorker Caption Contest"☆56Updated 4 months ago
- Implementation of the deepmind Flamingo vision-language model, based on Hugging Face language models and ready for training☆167Updated 2 years ago
- SlideVQA: A Dataset for Document Visual Question Answering on Multiple Images (AAAI2023)☆91Updated 3 months ago
- contrastive decoding☆202Updated 2 years ago
- Data and code for the paper "Inducing Positive Perspectives with Text Reframing"☆61Updated 2 years ago
- Open source code for AAAI 2023 Paper "BridgeTower: Building Bridges Between Encoders in Vision-Language Representation Learning"☆166Updated 2 years ago
- Code and model release for the paper "Task-aware Retrieval with Instructions" by Asai et al.☆163Updated last year
- Repository for research in the field of Responsible NLP at Meta.☆201Updated 2 months ago
- Hate-CLIPper: Multimodal Hateful Meme Classification with Explicit Cross-modal Interaction of CLIP features - Accepted at EMNLP 2022 Work…☆52Updated 3 months ago
- Chart-to-Text: Generating Natural Language Explanations for Charts by Adapting the Transformer Model☆156Updated last year
- ☆210Updated 3 months ago
- Resources for cultural NLP research☆98Updated 2 months ago
- Language Models Can See: Plugging Visual Controls in Text Generation☆257Updated 3 years ago
- Official code for paper "UniIR: Training and Benchmarking Universal Multimodal Information Retrievers" (ECCV 2024)☆155Updated 9 months ago
- MultiInstruct: Improving Multi-Modal Zero-Shot Learning via Instruction Tuning☆135Updated 2 years ago
- Code and data associated with the AmbiEnt dataset in "We're Afraid Language Models Aren't Modeling Ambiguity" (Liu et al., 2023)☆64Updated last year
- Data and code for the paper "NormBank: A Knowledge Bank of Situational Social Norms"☆29Updated 2 years ago
- Multilingual Large Language Models Evaluation Benchmark☆127Updated 10 months ago
- ☆16Updated 2 years ago
- ☆12Updated 3 years ago
- 🧀 Code and models for the ICML 2023 paper "Grounding Language Models to Images for Multimodal Inputs and Outputs".☆482Updated last year
- This is the official implementation of the paper "MM-SHAP: A Performance-agnostic Metric for Measuring Multimodal Contributions in Vision…☆30Updated last year
- Repository for Multilingual-VQA task created during HuggingFace JAX/Flax community week.☆34Updated 3 years ago
- Okapi: Instruction-tuned Large Language Models in Multiple Languages with Reinforcement Learning from Human Feedback☆97Updated last year
- NLX-GPT: A Model for Natural Language Explanations in Vision and Vision-Language Tasks, CVPR 2022 (Oral)☆48Updated last year
- [ICML 2023] Code for our paper “Compositional Exemplars for In-context Learning”.☆102Updated 2 years ago