xiaoachen98 / Open-LLaVA-NeXT
An open-source implementation for training LLaVA-NeXT.
☆395Updated 6 months ago
Alternatives and similar repositories for Open-LLaVA-NeXT
Users that are interested in Open-LLaVA-NeXT are comparing it to the libraries listed below
Sorting:
- The code for "TokenPacker: Efficient Visual Projector for Multimodal LLM".☆247Updated 4 months ago
- [ECCV 2024] Does Your Multi-modal LLM Truly See the Diagrams in Visual Math Problems?☆160Updated 2 weeks ago
- [ICLR 2025] Mathematical Visual Instruction Tuning for Multi-modal Large Language Models☆139Updated 5 months ago
- ☆383Updated 5 months ago
- (AAAI 2024) BLIVA: A Simple Multimodal LLM for Better Handling of Text-rich Visual Questions☆257Updated last year
- [ECCV2024] Grounded Multimodal Large Language Model with Localized Visual Tokenization☆563Updated 11 months ago
- A collection of multimodal reasoning papers, codes, datasets, benchmarks and resources.