SuperBruceJia / Awesome-Large-Vision-Language-ModelLinks
Awesome Large Vision-Language Model: A Curated List of Large Vision-Language Model
☆39Updated 5 months ago
Alternatives and similar repositories for Awesome-Large-Vision-Language-Model
Users that are interested in Awesome-Large-Vision-Language-Model are comparing it to the libraries listed below
Sorting:
- ☆78Updated 7 months ago
- Reading list for Multimodal Large Language Models☆69Updated 2 years ago
- MLLM-Bench: Evaluating Multimodal LLMs with Per-sample Criteria☆72Updated last year
- Official Implementation for EMNLP 2024 (main) "AgentReview: Exploring Academic Peer Review with LLM Agent."☆103Updated last year
- ☆140Updated 10 months ago
- Awesome Mixture of Experts (MoE): A Curated List of Mixture of Experts (MoE) and Mixture of Multimodal Experts (MoME)☆51Updated 3 months ago
- Collection of Tools and Papers related to Adapters / Parameter-Efficient Transfer Learning/ Fine-Tuning☆201Updated last year
- Residual Prompt Tuning: a method for faster and better prompt tuning.☆57Updated 2 years ago
- The code and data of We-Math, accepted by ACL 2025 main conference.☆134Updated last month
- Official PyTorch Implementation of MLLM Is a Strong Reranker: Advancing Multimodal Retrieval-augmented Generation via Knowledge-enhanced …☆90Updated last year
- [NeurIPS 2024] CharXiv: Charting Gaps in Realistic Chart Understanding in Multimodal LLMs☆138Updated 8 months ago
- Parameter-Efficient Fine-Tuning for Foundation Models☆105Updated 9 months ago
- [ACL 2024] ChartAssistant is a chart-based vision-language model for universal chart comprehension and reasoning.☆131Updated last year
- [NeurIPS 2024] A Novel Rank-Based Metric for Evaluating Large Language Models☆56Updated 7 months ago
- [TMLR 25] SFT or RL? An Early Investigation into Training R1-Like Reasoning Large Vision-Language Models☆146Updated 3 months ago
- Latest Advances on Reasoning of Multimodal Large Language Models (Multimodal R1 \ Visual R1) ) 🍓☆35Updated 9 months ago
- This project aims to collect and collate various datasets for multimodal large model training, including but not limited to pre-training …☆65Updated 8 months ago
- Offical Repository of "AtomThink: Multimodal Slow Thinking with Atomic Step Reasoning"☆60Updated last month
- [ICLR 2024] This is the repository for the paper titled "DePT: Decomposed Prompt Tuning for Parameter-Efficient Fine-tuning"☆101Updated last year
- ☆98Updated last year
- This is for ACL 2025 Findings Paper: From Specific-MLLMs to Omni-MLLMs: A Survey on MLLMs Aligned with Multi-modalitiesModels☆84Updated 2 weeks ago
- Automatically update arXiv papers about LLM Reasoning, LLM Evaluation, LLM & MLLM and Video Understanding using Github Actions.☆134Updated this week
- An open-source implementaion for fine-tuning Llama3.2-Vision series by Meta.☆174Updated 2 months ago
- [ICCV 2025] Auto Interpretation Pipeline and many other functionalities for Multimodal SAE Analysis.☆173Updated 3 months ago
- This repository provides a comprehensive collection of research papers focused on multimodal representation learning, all of which have b…☆83Updated 7 months ago
- ☆100Updated this week
- ☆33Updated last year
- Paper collections of multi-modal LLM for Math/STEM/Code.☆134Updated 2 months ago
- ☆58Updated 10 months ago
- ☆43Updated 7 months ago