lupantech / ScienceQA
Data and code for NeurIPS 2022 Paper "Learn to Explain: Multimodal Reasoning via Thought Chains for Science Question Answering".
☆643Updated 5 months ago
Alternatives and similar repositories for ScienceQA:
Users that are interested in ScienceQA are comparing it to the libraries listed below
- [NeurIPS 2023] Official implementations of "Cheap and Quick: Efficient Vision-Language Instruction Tuning for Large Language Models"☆515Updated last year
- ☆734Updated 9 months ago
- MathVista: data, code, and evaluation for Mathematical Reasoning in Visual Contexts☆282Updated 3 months ago
- Codes for "Chameleon: Plug-and-Play Compositional Reasoning with Large Language Models".☆1,119Updated last year
- MultimodalC4 is a multimodal extension of c4 that interleaves millions of images with text.☆918Updated 9 months ago
- MM-Vet: Evaluating Large Multimodal Models for Integrated Capabilities (ICML 2024)☆287Updated last month
- This repo contains evaluation code for the paper "MMMU: A Massive Multi-discipline Multimodal Understanding and Reasoning Benchmark for E…☆400Updated this week
- [COLM 2024] LoraHub: Efficient Cross-Task Generalization via Dynamic LoRA Composition☆617Updated 7 months ago
- Code/Data for the paper: "LLaVAR: Enhanced Visual Instruction Tuning for Text-Rich Image Understanding"☆265Updated 9 months ago
- Official repo for MM-REACT☆943Updated last year
- Reading list of Instruction-tuning. A trend starts from Natrural-Instruction (ACL 2022), FLAN (ICLR 2022) and T0 (ICLR 2022).☆765Updated last year
- [NIPS2023] RRHF & Wombat☆803Updated last year
- (CVPR2024)A benchmark for evaluating Multimodal LLMs using multiple-choice questions.☆332Updated last month
- Chatbot Arena meets multi-modality! Multi-Modality Arena allows you to benchmark vision-language models side-by-side while providing imag…☆504Updated 10 months ago
- Challenging BIG-Bench Tasks and Whether Chain-of-Thought Can Solve Them☆468Updated 8 months ago
- ☆767Updated 8 months ago
- [TLLM'23] PandaGPT: One Model To Instruction-Follow Them All☆782Updated last year
- Official implementation of paper "MiniGPT-5: Interleaved Vision-and-Language Generation via Generative Vokens"☆863Updated 3 months ago
- MMICL, a state-of-the-art VLM with the in context learning ability from ICL, PKU☆345Updated last year
- Codes for VPGTrans: Transfer Visual Prompt Generator across LLMs. VL-LLaMA, VL-Vicuna.☆271Updated last year
- PaL: Program-Aided Language Models (ICML 2023)☆482Updated last year
- OpenICL is an open-source framework to facilitate research, development, and prototyping of in-context learning.☆551Updated last year
- Inference-Time Intervention: Eliciting Truthful Answers from a Language Model☆510Updated last month
- Paper list about multimodal and large language models, only used to record papers I read in the daily arxiv for personal needs.☆605Updated this week
- Paper List for In-context Learning 🌷☆846Updated 5 months ago
- X-LLM: Bootstrapping Advanced Large Language Models by Treating Multi-Modalities as Foreign Languages☆309Updated last year
- 🐟 Code and models for the NeurIPS 2023 paper "Generating Images with Multimodal Language Models".☆448Updated last year
- GPT4Tools is an intelligent system that can automatically decide, control, and utilize different visual foundation models, allowing the u…☆768Updated last year
- Aligning LMMs with Factually Augmented RLHF☆352Updated last year
- LLaVA-Plus: Large Language and Vision Assistants that Plug and Learn to Use Skills☆726Updated last year