opendatalab / CHARM
[ACL 2024 Main Conference] Chinese commonsense benchmark for LLMs
☆23Updated 3 months ago
Related projects ⓘ
Alternatives and complementary repositories for CHARM
- AAAI 2024: Visual Instruction Generation and Correction☆90Updated 9 months ago
- ☆31Updated 4 months ago
- This is the repo for the paper Multi-Agent Collaborative Data Selection for Efficient LLM Pretraining.☆27Updated 2 weeks ago
- A Fine-Grained Instruction Tuning Dataset and Model for Remote Sensing Vision-Language Understanding☆63Updated 3 months ago
- Pink: Unveiling the Power of Referential Comprehension for Multi-modal LLMs☆76Updated 4 months ago
- 【ArXiv】PDF-Wukong: A Large Multimodal Model for Efficient Long PDF Reading with End-to-End Sparse Sampling☆93Updated 3 weeks ago
- ☆33Updated 4 months ago
- ☆28Updated last week
- Accompanying repo for 'Charting New Territories: Exploring the Geographic and Geospatial Capabilities of Multimodal LLMs' project☆24Updated 2 months ago
- VGI-Enhanced multimodal large language model for remote sensing images.☆104Updated last month
- VHM: Versatile and Honest Vision Language Model for Remote Sensing Image Analysis☆46Updated 7 months ago
- ☆78Updated 9 months ago
- Draw-and-Understand: Leveraging Visual Prompts to Enable MLLMs to Comprehend What You Want☆59Updated 2 weeks ago
- [CVPR 2024] Official Code for the Paper "Compositional Chain-of-Thought Prompting for Large Multimodal Models"☆78Updated 4 months ago
- The proposed simulated dataset consisting of 9,536 charts and associated data annotations in CSV format.☆21Updated 8 months ago
- The official implementation of the paper “LOKI:A Comprehensive Synthetic Data Detection Benchmark using Large Multimodal Models”☆107Updated 2 weeks ago
- ☆67Updated 6 months ago
- [NeurIPS2024] Repo for the paper `ControlMLLM: Training-Free Visual Prompt Learning for Multimodal Large Language Models'☆88Updated last month
- ☆84Updated 11 months ago
- ☆94Updated last year
- ☆14Updated this week
- [NeurIPS 2024] Needle In A Multimodal Haystack (MM-NIAH): A comprehensive benchmark designed to systematically evaluate the capability of…☆97Updated 2 weeks ago
- This is the official repo for Contrastive Vision-Language Alignment Makes Efficient Instruction Learner.☆20Updated 11 months ago
- ✨✨ MME-RealWorld: Could Your Multimodal LLM Challenge High-Resolution Real-World Scenarios that are Difficult for Humans?☆77Updated last month
- [ECCV 2024] Paying More Attention to Image: A Training-Free Method for Alleviating Hallucination in LVLMs☆64Updated this week
- [NeurIPS 2024] MoVA: Adapting Mixture of Vision Experts to Multimodal Context☆131Updated last month
- Official Repo of "MMBench: Is Your Multi-modal Model an All-around Player?"☆163Updated 2 months ago
- SVIT: Scaling up Visual Instruction Tuning☆163Updated 4 months ago
- The official repo for “TextCoT: Zoom In for Enhanced Multimodal Text-Rich Image Understanding”.☆32Updated last month
- [CVPR 2024] LION: Empowering Multimodal Large Language Model with Dual-Level Visual Knowledge☆122Updated 3 months ago