patrick-tssn / Awesome-Colorful-LLMLinks
Recent advancements propelled by large language models (LLMs), encompassing an array of domains including Vision, Audio, Agent, Robotics, Fundamental Sciences such as Mathematics, and Ominous.
☆122Updated this week
Alternatives and similar repositories for Awesome-Colorful-LLM
Users that are interested in Awesome-Colorful-LLM are comparing it to the libraries listed below
Sorting:
- [ACL 2024] PCA-Bench: Evaluating Multimodal Large Language Models in Perception-Cognition-Action Chain☆104Updated last year
- Touchstone: Evaluating Vision-Language Models by Language Models☆83Updated last year
- ☆63Updated last year
- Harnessing 1.4M GPT4V-synthesized Data for A Lite Vision-Language Model☆261Updated 11 months ago
- The official repository for "2.5 Years in Class: A Multimodal Textbook for Vision-Language Pretraining"☆156Updated 2 months ago
- Official code of *Virgo: A Preliminary Exploration on Reproducing o1-like MLLM*☆103Updated this week
- [NeurIPS 2024] Needle In A Multimodal Haystack (MM-NIAH): A comprehensive benchmark designed to systematically evaluate the capability of…☆117Updated 6 months ago
- [CVPR'24] RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback☆278Updated 8 months ago
- ControlLLM: Augment Language Models with Tools by Searching on Graphs☆191Updated 10 months ago
- MMICL, a state-of-the-art VLM with the in context learning ability from ICL, PKU☆350Updated last year
- A RLHF Infrastructure for Vision-Language Models☆177Updated 6 months ago
- Official repository of MMDU dataset☆91Updated 8 months ago
- ☆73Updated last year
- MM-Vet: Evaluating Large Multimodal Models for Integrated Capabilities (ICML 2024)☆301Updated 4 months ago
- [ICLR'24] Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuning☆279Updated last year
- [TMLR] Public code repo for paper "A Single Transformer for Scalable Vision-Language Modeling"☆139Updated 6 months ago
- ✨✨R1-Reward: Training Multimodal Reward Model Through Stable Reinforcement Learning☆133Updated 3 weeks ago
- ☆99Updated last year
- RM-R1: Unleashing the Reasoning Potential of Reward Models☆93Updated this week
- ☆147Updated 7 months ago
- [ACL 2024 Findings] "TempCompass: Do Video LLMs Really Understand Videos?", Yuanxin Liu, Shicheng Li, Yi Liu, Yuxiang Wang, Shuhuai Ren, …☆115Updated last month
- Official code for paper "UniIR: Training and Benchmarking Universal Multimodal Information Retrievers" (ECCV 2024)☆149Updated 8 months ago
- An Easy-to-use, Scalable and High-performance RLHF Framework designed for Multimodal Models.☆125Updated last month
- A Self-Training Framework for Vision-Language Reasoning☆80Updated 4 months ago
- A curated list of the papers, repositories, tutorials, and anythings related to the large language models for tools☆67Updated last year
- The official GitHub page for ''Evaluating Object Hallucination in Large Vision-Language Models''☆207Updated last year
- Code for Math-LLaVA: Bootstrapping Mathematical Reasoning for Multimodal Large Language Models☆84Updated 11 months ago
- Official code for Paper "Mantis: Multi-Image Instruction Tuning" [TMLR2024]☆215Updated 2 months ago
- An benchmark for evaluating the capabilities of large vision-language models (LVLMs)☆45Updated last year
- ☆74Updated 11 months ago