princeton-nlp / PTP
Improving Language Understanding from Screenshots. Paper: https://arxiv.org/abs/2402.14073
☆28Updated 9 months ago
Alternatives and similar repositories for PTP:
Users that are interested in PTP are comparing it to the libraries listed below
- This repository contains the code and data for the paper "VisOnlyQA: Large Vision Language Models Still Struggle with Visual Perception o…☆22Updated 2 weeks ago
- Evaluation framework for paper "VisualWebBench: How Far Have Multimodal LLMs Evolved in Web Page Understanding and Grounding?"☆54Updated 5 months ago
- ☆59Updated 7 months ago
- MUFFIN: Curating Multi-Faceted Instructions for Improving Instruction-Following☆16Updated 5 months ago
- [NeurIPS 2024] OlympicArena: Benchmarking Multi-discipline Cognitive Reasoning for Superintelligent AI☆99Updated last month
- Unofficial Implementation of Chain-of-Thought Reasoning Without Prompting☆32Updated last year
- The code and data for the paper JiuZhang3.0☆43Updated 10 months ago
- ☆10Updated 10 months ago
- This is the repo for our paper "Mr-Ben: A Comprehensive Meta-Reasoning Benchmark for Large Language Models"☆47Updated 5 months ago
- Code release for "SPIQA: A Dataset for Multimodal Question Answering on Scientific Papers" [NeurIPS D&B, 2024]☆57Updated 3 months ago
- Visual and Embodied Concepts evaluation benchmark☆21Updated last year
- ☆15Updated last year
- Code for Math-LLaVA: Bootstrapping Mathematical Reasoning for Multimodal Large Language Models☆83Updated 9 months ago
- ☆49Updated last year
- [ICLR 2025] LongPO: Long Context Self-Evolution of Large Language Models through Short-to-Long Preference Optimization☆31Updated last month
- M-STAR (Multimodal Self-Evolving TrAining for Reasoning) Project. Diving into Self-Evolving Training for Multimodal Reasoning☆56Updated 3 months ago
- [ICLR'24 spotlight] Tool-Augmented Reward Modeling☆47Updated 3 months ago
- [NAACL 2025] Source code for MMEvalPro, a more trustworthy and efficient benchmark for evaluating LMMs☆23Updated 6 months ago
- Large Language Models Can Self-Improve in Long-context Reasoning☆68Updated 4 months ago
- The rule-based evaluation subset and code implementation of Omni-MATH☆19Updated 3 months ago
- Code for the arXiv preprint "The Unreasonable Effectiveness of Easy Training Data"☆47Updated last year
- Official repository for paper "Weak-to-Strong Extrapolation Expedites Alignment"☆74Updated 10 months ago
- Code and data for the ACL 2024 Findings paper "Do LVLMs Understand Charts? Analyzing and Correcting Factual Errors in Chart Captioning"☆25Updated 10 months ago
- Official Code of IdealGPT☆34Updated last year
- [ACL 2024 Findings] CriticBench: Benchmarking LLMs for Critique-Correct Reasoning☆24Updated last year
- This repo contains evaluation code for the paper "MileBench: Benchmarking MLLMs in Long Context"☆31Updated 9 months ago
- Evaluate the Quality of Critique☆34Updated 10 months ago
- GSM-Plus: Data, Code, and Evaluation for Enhancing Robust Mathematical Reasoning in Math Word Problems.☆59Updated 9 months ago
- ☆34Updated last year
- The source code for running LLMs on the AAAR-1.0 benchmark.☆16Updated last week