ZhangYiqun018 / StickerConvLinks
☆55Updated last year
Alternatives and similar repositories for StickerConv
Users that are interested in StickerConv are comparing it to the libraries listed below
Sorting:
- Official repo for "AlignGPT: Multi-modal Large Language Models with Adaptive Alignment Capability"☆33Updated last year
- ☆82Updated last year
- [EMNLP 2024] ”ESC-Eval: Evaluating Emotion Support Conversations in Large Language Models“☆19Updated last year
- ☆108Updated last week
- (ICLR'25) A Comprehensive Framework for Developing and Evaluating Multimodal Role-Playing Agents☆86Updated 8 months ago
- [ACL'2024 Findings] GAOKAO-MM: A Chinese Human-Level Benchmark for Multimodal Models Evaluation☆67Updated last year
- [ICLR 2025] ChartMimic: Evaluating LMM’s Cross-Modal Reasoning Capability via Chart-to-Code Generation☆123Updated 3 months ago
- 自己阅读的多模态对话系统论文(及部分笔记)汇总☆23Updated 2 years ago
- Latest Advances on Reasoning of Multimodal Large Language Models (Multimodal R1 \ Visual R1) ) 🍓☆34Updated 6 months ago
- ☆27Updated 11 months ago
- ☆17Updated 6 months ago
- This repository contains the code for SFT, RLHF, and DPO, designed for vision-based LLMs, including the LLaVA models and the LLaMA-3.2-vi…☆114Updated 3 months ago
- ☆14Updated last year
- ☆83Updated last year
- mPLUG-HalOwl: Multimodal Hallucination Evaluation and Mitigating☆98Updated last year
- RoleInteract: Evaluating the Social Interaction of Role-Playing Agents☆61Updated 11 months ago
- MoCLE (First MLLM with MoE for instruction customization and generalization!) (https://arxiv.org/abs/2312.12379)☆43Updated 3 months ago
- Paper, dataset and code list for multimodal dialogue.☆22Updated 9 months ago
- A Self-Training Framework for Vision-Language Reasoning☆86Updated 8 months ago
- [ICML'2024] Can AI Assistants Know What They Don't Know?☆83Updated last year
- [SIGIR'24] The official implementation code of MOELoRA.☆182Updated last year
- This repository will continuously update the latest papers, technical reports, benchmarks about multimodal reasoning!☆51Updated 6 months ago
- The demo, code and data of FollowRAG☆75Updated 3 months ago
- Repository for Label Words are Anchors: An Information Flow Perspective for Understanding In-Context Learning☆165Updated last year
- code for "CoMT: A Novel Benchmark for Chain of Multi-modal Thought on Large Vision-Language Models"☆19Updated 6 months ago
- Data for evaluating GPT-4V☆11Updated last year
- [ACL24] EmoBench: Evaluating the Emotional Intelligence of Large Language Models☆92Updated 4 months ago
- mPLUG: Effective and Efficient Vision-Language Learning by Cross-modal Skip-connections. (EMNLP 2022)☆97Updated 2 years ago
- Code and model for AAAI 2024: UMIE: Unified Multimodal Information Extraction with Instruction Tuning☆40Updated last year
- [ACM MM 2022 Oral] This is the official implementation of "SER30K: A Large-Scale Dataset for Sticker Emotion Recognition"☆25Updated 2 years ago