will-singularity / Skywork-MM
Empirical Study Towards Building An Effective Multi-Modal Large Language Model
☆23Updated last year
Alternatives and similar repositories for Skywork-MM:
Users that are interested in Skywork-MM are comparing it to the libraries listed below
- Touchstone: Evaluating Vision-Language Models by Language Models☆82Updated last year
- A Framework for Decoupling and Assessing the Capabilities of VLMs☆40Updated 8 months ago
- ☆17Updated last year
- ☆20Updated last year
- Web2Code: A Large-scale Webpage-to-Code Dataset and Evaluation Framework for Multimodal LLMs☆75Updated 4 months ago
- LMM solved catastrophic forgetting, AAAI2025☆39Updated 4 months ago
- ☆86Updated 8 months ago
- ☆73Updated last year
- Our 2nd-gen LMM☆33Updated 9 months ago
- ☆54Updated 7 months ago
- MLLM-DataEngine: An Iterative Refinement Approach for MLLM☆44Updated 9 months ago
- Exploring Efficient Fine-Grained Perception of Multimodal Large Language Models☆60Updated 4 months ago
- MM-Instruct: Generated Visual Instructions for Large Multimodal Model Alignment☆33Updated 8 months ago
- ☆29Updated 7 months ago
- The code for "VISTA: Enhancing Long-Duration and High-Resolution Video Understanding by VIdeo SpatioTemporal Augmentation" [CVPR2025]☆14Updated 3 weeks ago
- Synthetic data generation pipelines for text-rich images.☆45Updated 2 weeks ago
- Offical Code for GPT4Video: A Unified Multimodal Large Language Model for lnstruction-Followed Understanding and Safety-Aware Generation☆138Updated 4 months ago
- [NeurIPS 2024] Needle In A Multimodal Haystack (MM-NIAH): A comprehensive benchmark designed to systematically evaluate the capability of…☆113Updated 3 months ago
- Official repository of MMDU dataset☆86Updated 5 months ago
- [ACL 2024] PCA-Bench: Evaluating Multimodal Large Language Models in Perception-Cognition-Action Chain☆102Updated last year
- This repo contains the code for "MEGA-Bench Scaling Multimodal Evaluation to over 500 Real-World Tasks" [ICLR2025]☆60Updated last week
- [ArXiv] V2PE: Improving Multimodal Long-Context Capability of Vision-Language Models with Variable Visual Position Encoding☆31Updated 3 months ago
- LLaVA combines with Magvit Image tokenizer, training MLLM without an Vision Encoder. Unifying image understanding and generation.☆35Updated 9 months ago
- ☆61Updated last year
- ☆49Updated last year
- MM-Interleaved: Interleaved Image-Text Generative Modeling via Multi-modal Feature Synchronizer☆220Updated 11 months ago
- [NeurIPS-24] This is the official implementation of the paper "DeepStack: Deeply Stacking Visual Tokens is Surprisingly Simple and Effect…☆35Updated 9 months ago
- Official code of *Virgo: A Preliminary Exploration on Reproducing o1-like MLLM*☆96Updated 3 weeks ago
- Official Repository of VideoLLaMB: Long Video Understanding with Recurrent Memory Bridges☆65Updated 3 weeks ago
- ☆133Updated last year