thuiar / MMLALinks
The first comprehensive multimodal language analysis benchmark for evaluating foundation models
☆24Updated last month
Alternatives and similar repositories for MMLA
Users that are interested in MMLA are comparing it to the libraries listed below
Sorting:
- ☆58Updated last year
 - MIntRec2.0 is the first large-scale dataset for multimodal intent recognition and out-of-scope detection in multi-party conversations (IC…☆66Updated 2 months ago
 - (NIPS 2025) OpenOmni: Official implementation of Advancing Open-Source Omnimodal Large Language Models with Progressive Multimodal Align…☆107Updated last month
 - ☆21Updated 9 months ago
 - HumanOmni☆201Updated 7 months ago
 - Official repo for "AlignGPT: Multi-modal Large Language Models with Adaptive Alignment Capability"☆33Updated last year
 - Synth-Empathy: Towards High-Quality Synthetic Empathy Data☆16Updated 8 months ago
 - Code and resources for the NeurIPS 2025 Paper "BMMR: A Large-Scale Bilingual Multimodal Multi-Discipline Reasoning Dataset" by Zhiheng X…☆15Updated 2 weeks ago
 - EchoInk-R1: Exploring Audio-Visual Reasoning in Multimodal LLMs via Reinforcement Learning [🔥The Exploration of R1 for General Audio-Vi…☆60Updated 5 months ago
 - Explainable Multimodal Emotion Reasoning (EMER), OV-MER (ICML), and AffectGPT (ICML, Oral)☆271Updated 2 months ago
 - Multimodal Empathetic Chatbot☆51Updated last year
 - WorldSense: Evaluating Real-world Omnimodal Understanding for Multimodal LLMs☆31Updated last month
 - ☆19Updated 4 months ago
 - GPT-4V with Emotion☆95Updated last year
 - ☆14Updated 10 months ago
 - code for "CoMT: A Novel Benchmark for Chain of Multi-modal Thought on Large Vision-Language Models"☆19Updated 7 months ago
 - ☆116Updated 2 weeks ago
 - (ICLR'25) A Comprehensive Framework for Developing and Evaluating Multimodal Role-Playing Agents☆87Updated 9 months ago
 - This repository hosts the code, data and model weight of PanoSent.☆56Updated 3 months ago
 - The code and data of We-Math, accepted by ACL 2025 main conference.☆133Updated 2 weeks ago
 - A project for tri-modal LLM benchmarking and instruction tuning.☆48Updated 7 months ago
 - [ICML'25 Spotlight] Catch Your Emotion: Sharpening Emotion Perception in Multimodal Large Language Models☆39Updated last month
 - On Path to Multimodal Generalist: General-Level and General-Bench☆19Updated 3 months ago
 - [NeurIPS 2025] More Thinking, Less Seeing? Assessing Amplified Hallucination in Multimodal Reasoning Models☆65Updated 5 months ago
 - Sparrow: Data-Efficient Video-LLM with Text-to-Image Augmentation☆30Updated 7 months ago
 - ☆84Updated last year
 - The official implement of paper 《DaMo: Data Mixing Optimizer in Fine-tuning Multimodal LLMs for Mobile Phone Agents》☆28Updated last week
 - MIntRec: A New Dataset for Multimodal Intent Recognition (ACM MM 2022)☆115Updated 6 months ago
 - [ACL24] EmoBench: Evaluating the Emotional Intelligence of Large Language Models☆97Updated 5 months ago
 - [ACL 2025 (Findings)] DEMO: Reframing Dialogue Interaction with Fine-grained Element Modeling☆20Updated 10 months ago