zchoi / Multi-Modal-Large-Language-LearningLinks
Awesome multi-modal large language paper/project, collections of popular training strategies, e.g., PEFT, LoRA.
β27Updated 11 months ago
Alternatives and similar repositories for Multi-Modal-Large-Language-Learning
Users that are interested in Multi-Modal-Large-Language-Learning are comparing it to the libraries listed below
Sorting:
- π₯ Omni large models and datasets for understanding and generating multi-modalities.β15Updated 8 months ago
- [EMNLP'23] The official GitHub page for ''Evaluating Object Hallucination in Large Vision-Language Models''β85Updated last year
- Official Pytorch implementation of "Interpreting and Editing Vision-Language Representations to Mitigate Hallucinations" (ICLR '25)β75Updated last month
- [Arxiv] Aligning Modalities in Vision Large Language Models via Preference Fine-tuningβ86Updated last year
- A Survey on Benchmarks of Multimodal Large Language Modelsβ119Updated 2 weeks ago
- [ECCV 2024] BenchLMM: Benchmarking Cross-style Visual Capability of Large Multimodal Modelsβ85Updated 10 months ago
- Enhancing Large Vision Language Models with Self-Training on Image Comprehension.β68Updated last year
- An LLM-free Multi-dimensional Benchmark for Multi-modal Hallucination Evaluationβ126Updated last year