InternScience / SciEvalKitLinks
A unified evaluation toolkit and leaderboard for rigorously assessing the scientific intelligence of large language and vision–language models across the full research workflow.
☆69Updated this week
Alternatives and similar repositories for SciEvalKit
Users that are interested in SciEvalKit are comparing it to the libraries listed below
Sorting:
- ☆61Updated last month
- Probing Scientific General Intelligence of LLMs with Scientist-Aligned Workflows☆147Updated 2 weeks ago
- Official codes of "Monet: Reasoning in Latent Visual Space Beyond Image and Language"☆123Updated last month
- [CVPR2025 Highlight] Insight-V: Exploring Long-Chain Visual Reasoning with Multimodal Large Language Models☆233Updated 2 months ago
- Machine Mental Imagery: Empower Multimodal Reasoning with Latent Visual Tokens (arXiv 2025)☆238Updated 6 months ago
- Pixel-Level Reasoning Model trained with RL [NeuIPS25]☆273Updated 3 months ago
- [NeurIPS 2025] Reinforcing Spatial Reasoning in Vision-Language Models with Interwoven Thinking and Visual Drawing☆90Updated 6 months ago
- Interleaving Reasoning: Next-Generation Reasoning Systems for AGI☆250Updated 3 months ago
- Official repo for "PAPO: Perception-Aware Policy Optimization for Multimodal Reasoning"☆111Updated last week
- OpenVLThinker: An Early Exploration to Vision-Language Reasoning via Iterative Self-Improvement☆129Updated 6 months ago
- Codes for Visual Sketchpad: Sketching as a Visual Chain of Thought for Multimodal Language Models☆277Updated 6 months ago
- Data and Code for CVPR 2025 paper "MMVU: Measuring Expert-Level Multi-Discipline Video Understanding"☆77Updated 11 months ago
- ☆60Updated 2 months ago
- [ICCV25 Oral] Token Activation Map to Visually Explain Multimodal LLMs☆166Updated last month
- Official Code for "Mini-o3: Scaling Up Reasoning Patterns and Interaction Turns for Visual Search"☆395Updated last week
- A collection of awesome think with videos papers.☆86Updated 2 months ago
- This repo is the official implementation of "Euclid’s Gift: Enhancing Spatial Perception and Reasoning in Vision‑Language Models via Geom…☆26Updated 2 months ago
- ☆46Updated last year
- Draw ALL Your Imagine: A Holistic Benchmark and Agent Framework for Complex Instruction-based Image Generation☆23Updated 4 months ago
- OpenMMReasoner: Pushing the Frontiers for Multimodal Reasoning with an Open and General Recipe☆141Updated last month
- https://huggingface.co/datasets/multimodal-reasoning-lab/Zebra-CoT☆117Updated last week
- [TMLR 25] SFT or RL? An Early Investigation into Training R1-Like Reasoning Large Vision-Language Models☆149Updated 3 months ago
- MLLM-Tool: A Multimodal Large Language Model For Tool Agent Learning☆138Updated 3 months ago
- PyTorch implementation of NEPA☆308Updated last week
- Official code for NeurIPS 2025 paper "GRIT: Teaching MLLMs to Think with Images"☆173Updated 3 weeks ago
- [NeurIPS 2025] NoisyRollout: Reinforcing Visual Reasoning with Data Augmentation☆104Updated 4 months ago
- This repository is the official implementation of "Look-Back: Implicit Visual Re-focusing in MLLM Reasoning".☆83Updated 6 months ago
- [NeurIPS2024] Repo for the paper `ControlMLLM: Training-Free Visual Prompt Learning for Multimodal Large Language Models'☆205Updated 6 months ago
- Dimple, the first Discrete Diffusion Multimodal Large Language Model☆114Updated 6 months ago
- Self Evolving Large Multimodal Models with Continuous Rewards☆19Updated 2 months ago